2026-03-10T12:27:33.262 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-10T12:27:33.267 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T12:27:33.287 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028 branch: squid description: orch/cephadm/workunits/{0-distro/ubuntu_22.04 agent/on mon_election/connectivity task/test_extra_daemon_features} email: null first_in_suite: false flavor: default job_id: '1028' ktype: distro last_in_suite: false machine_type: vps name: kyr-2026-03-10_01:00:38-orch-squid-none-default-vps no_nested_subset: false os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 3 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: true mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath workunit: branch: tt-squid sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 seed: 8043 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b targets: vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIm90GeTTyY024yP1tRM6qvVajJOlH5oQHLgmTj4fzd/4O2ZIG/E/IS2D3Qg07FwkFA6JPL1rO5i9mWm6jhpa+o= vm05.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBI4bDeQ6PrctXLf0y3KxW2GiR+3XxkEhRUHFlwDLAALsrwmXQsAsQEBQXbu0Al2QT6p7uqt5rc5KBWX1dnmV65s= tasks: - install: null - cephadm: null - exec: all-hosts: - mkdir /etc/cephadm_testing - cephadm.apply: specs: - extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon - custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest - custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z - cephadm.wait_for_service: service: mon - cephadm.wait_for_service: service: container.foo - cephadm.wait_for_service: service: container.bar - exec: host.a: - 'set -ex FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) sleep 60 # check extra container and entrypoint args written to mon unit run file grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run # check that custom container properly wrote content to file. # This requires the custom config, extra container args, and # entrypoint args to all be working in order for this to have # been written. The container entrypoint was set up with custom_configs, # the content and where to write to with the entrypoint args, and the mounting # of the /etc/cephadm_testing dir with extra container args grep "testing_custom_containers" /etc/cephadm_testing/testing.txt # Verify that container bar''s init containers and primary container # ran successfully dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name ''container.bar.*'') test -n "$dir" grep ok ${dir}/data/primary.txt grep from=centos ${dir}/data/from.txt test -s ${dir}/data/presized.dat ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-10_01:00:38 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-10T12:27:33.287 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa; will attempt to use it 2026-03-10T12:27:33.287 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa/tasks 2026-03-10T12:27:33.287 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-10T12:27:33.287 INFO:teuthology.task.internal:Checking packages... 2026-03-10T12:27:33.288 INFO:teuthology.task.internal:Checking packages for os_type 'ubuntu', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-10T12:27:33.288 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-10T12:27:33.288 INFO:teuthology.packaging:ref: None 2026-03-10T12:27:33.288 INFO:teuthology.packaging:tag: None 2026-03-10T12:27:33.288 INFO:teuthology.packaging:branch: squid 2026-03-10T12:27:33.288 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:27:33.288 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&ref=squid 2026-03-10T12:27:33.916 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:27:33.917 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-10T12:27:33.918 INFO:teuthology.task.internal:no buildpackages task found 2026-03-10T12:27:33.918 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-10T12:27:33.918 INFO:teuthology.task.internal:Saving configuration 2026-03-10T12:27:33.923 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-10T12:27:33.924 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-10T12:27:33.930 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm01.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 12:26:30.134200', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:01', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIm90GeTTyY024yP1tRM6qvVajJOlH5oQHLgmTj4fzd/4O2ZIG/E/IS2D3Qg07FwkFA6JPL1rO5i9mWm6jhpa+o='} 2026-03-10T12:27:33.935 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm05.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 12:26:30.134607', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:05', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBI4bDeQ6PrctXLf0y3KxW2GiR+3XxkEhRUHFlwDLAALsrwmXQsAsQEBQXbu0Al2QT6p7uqt5rc5KBWX1dnmV65s='} 2026-03-10T12:27:33.935 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-10T12:27:33.936 INFO:teuthology.task.internal:roles: ubuntu@vm01.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0'] 2026-03-10T12:27:33.936 INFO:teuthology.task.internal:roles: ubuntu@vm05.local - ['host.b', 'mon.b', 'mgr.b', 'osd.1'] 2026-03-10T12:27:33.936 INFO:teuthology.run_tasks:Running task console_log... 2026-03-10T12:27:33.941 DEBUG:teuthology.task.console_log:vm01 does not support IPMI; excluding 2026-03-10T12:27:33.946 DEBUG:teuthology.task.console_log:vm05 does not support IPMI; excluding 2026-03-10T12:27:33.947 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f8af56a7eb0>, signals=[15]) 2026-03-10T12:27:33.947 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-10T12:27:33.947 INFO:teuthology.task.internal:Opening connections... 2026-03-10T12:27:33.947 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-10T12:27:33.948 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:27:34.005 DEBUG:teuthology.task.internal:connecting to ubuntu@vm05.local 2026-03-10T12:27:34.005 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:27:34.063 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-10T12:27:34.064 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-10T12:27:34.075 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-10T12:27:34.075 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:NAME="Ubuntu" 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="22.04" 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_CODENAME=jammy 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:ID=ubuntu 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE=debian 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-10T12:27:34.121 INFO:teuthology.orchestra.run.vm01.stdout:UBUNTU_CODENAME=jammy 2026-03-10T12:27:34.121 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-10T12:27:34.126 DEBUG:teuthology.orchestra.run.vm05:> uname -m 2026-03-10T12:27:34.128 INFO:teuthology.orchestra.run.vm05.stdout:x86_64 2026-03-10T12:27:34.128 DEBUG:teuthology.orchestra.run.vm05:> cat /etc/os-release 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:NAME="Ubuntu" 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:VERSION_ID="22.04" 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:VERSION_CODENAME=jammy 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:ID=ubuntu 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:ID_LIKE=debian 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-10T12:27:34.171 INFO:teuthology.orchestra.run.vm05.stdout:UBUNTU_CODENAME=jammy 2026-03-10T12:27:34.171 INFO:teuthology.lock.ops:Updating vm05.local on lock server 2026-03-10T12:27:34.175 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-10T12:27:34.177 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-10T12:27:34.178 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-10T12:27:34.178 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-10T12:27:34.179 DEBUG:teuthology.orchestra.run.vm05:> test '!' -e /home/ubuntu/cephtest 2026-03-10T12:27:34.215 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-10T12:27:34.215 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-10T12:27:34.216 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-10T12:27:34.223 DEBUG:teuthology.orchestra.run.vm05:> test -z $(ls -A /var/lib/ceph) 2026-03-10T12:27:34.225 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T12:27:34.259 INFO:teuthology.orchestra.run.vm05.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T12:27:34.259 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-10T12:27:34.266 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-10T12:27:34.268 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:27:34.499 DEBUG:teuthology.orchestra.run.vm05:> test -e /ceph-qa-ready 2026-03-10T12:27:34.502 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:27:34.715 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-10T12:27:34.716 INFO:teuthology.task.internal:Creating test directory... 2026-03-10T12:27:34.716 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T12:27:34.717 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T12:27:34.720 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-10T12:27:34.721 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-10T12:27:34.722 INFO:teuthology.task.internal:Creating archive directory... 2026-03-10T12:27:34.722 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T12:27:34.763 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T12:27:34.768 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-10T12:27:34.769 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-10T12:27:34.769 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T12:27:34.808 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:27:34.809 DEBUG:teuthology.orchestra.run.vm05:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T12:27:34.811 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:27:34.811 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T12:27:34.851 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T12:27:34.857 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:27:34.859 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:27:34.861 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:27:34.863 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:27:34.864 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-10T12:27:34.865 INFO:teuthology.task.internal:Configuring sudo... 2026-03-10T12:27:34.865 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T12:27:34.907 DEBUG:teuthology.orchestra.run.vm05:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T12:27:34.914 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-10T12:27:34.916 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-10T12:27:34.916 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T12:27:34.955 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T12:27:34.959 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T12:27:35.001 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T12:27:35.045 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:27:35.045 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T12:27:35.094 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T12:27:35.097 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T12:27:35.142 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:27:35.142 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T12:27:35.191 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-10T12:27:35.192 DEBUG:teuthology.orchestra.run.vm05:> sudo service rsyslog restart 2026-03-10T12:27:35.246 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-10T12:27:35.248 INFO:teuthology.task.internal:Starting timer... 2026-03-10T12:27:35.248 INFO:teuthology.run_tasks:Running task pcp... 2026-03-10T12:27:35.250 INFO:teuthology.run_tasks:Running task selinux... 2026-03-10T12:27:35.252 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-10T12:27:35.252 INFO:teuthology.task.selinux:Excluding vm05: VMs are not yet supported 2026-03-10T12:27:35.252 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-10T12:27:35.252 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-10T12:27:35.252 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-10T12:27:35.252 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-10T12:27:35.254 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-10T12:27:35.254 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-10T12:27:35.255 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-10T12:27:36.028 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-10T12:27:36.033 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-10T12:27:36.033 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventory6kstih37 --limit vm01.local,vm05.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-10T12:29:44.363 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm01.local'), Remote(name='ubuntu@vm05.local')] 2026-03-10T12:29:44.363 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-10T12:29:44.364 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:29:44.429 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-10T12:29:44.640 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-10T12:29:44.641 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm05.local' 2026-03-10T12:29:44.641 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:29:44.706 DEBUG:teuthology.orchestra.run.vm05:> true 2026-03-10T12:29:44.916 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm05.local' 2026-03-10T12:29:44.916 INFO:teuthology.run_tasks:Running task clock... 2026-03-10T12:29:44.919 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-10T12:29:44.919 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T12:29:44.920 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T12:29:44.921 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T12:29:44.921 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T12:29:44.937 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-10T12:29:44.937 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Command line: ntpd -gq 2026-03-10T12:29:44.937 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: ---------------------------------------------------- 2026-03-10T12:29:44.937 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: ntp-4 is maintained by Network Time Foundation, 2026-03-10T12:29:44.937 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-10T12:29:44.937 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: corporation. Support and training for ntp-4 are 2026-03-10T12:29:44.937 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: available at https://www.nwtime.org/support 2026-03-10T12:29:44.937 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: ---------------------------------------------------- 2026-03-10T12:29:44.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: proto: precision = 0.029 usec (-25) 2026-03-10T12:29:44.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: basedate set to 2022-02-04 2026-03-10T12:29:44.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: gps base set to 2022-02-06 (week 2196) 2026-03-10T12:29:44.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-10T12:29:44.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-10T12:29:44.939 INFO:teuthology.orchestra.run.vm01.stderr:10 Mar 12:29:44 ntpd[16092]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 73 days ago 2026-03-10T12:29:44.939 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Listen and drop on 0 v6wildcard [::]:123 2026-03-10T12:29:44.939 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-10T12:29:44.940 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Listen normally on 2 lo 127.0.0.1:123 2026-03-10T12:29:44.940 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Listen normally on 3 ens3 192.168.123.101:123 2026-03-10T12:29:44.940 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Listen normally on 4 lo [::1]:123 2026-03-10T12:29:44.940 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:1%2]:123 2026-03-10T12:29:44.940 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:44 ntpd[16092]: Listening on routing socket on fd #22 for interface updates 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Command line: ntpd -gq 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: ---------------------------------------------------- 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: ntp-4 is maintained by Network Time Foundation, 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: corporation. Support and training for ntp-4 are 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: available at https://www.nwtime.org/support 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: ---------------------------------------------------- 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: proto: precision = 0.031 usec (-25) 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: basedate set to 2022-02-04 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: gps base set to 2022-02-06 (week 2196) 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-10T12:29:44.976 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-10T12:29:44.977 INFO:teuthology.orchestra.run.vm05.stderr:10 Mar 12:29:44 ntpd[16107]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 73 days ago 2026-03-10T12:29:44.977 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Listen and drop on 0 v6wildcard [::]:123 2026-03-10T12:29:44.977 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-10T12:29:44.977 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Listen normally on 2 lo 127.0.0.1:123 2026-03-10T12:29:44.977 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Listen normally on 3 ens3 192.168.123.105:123 2026-03-10T12:29:44.977 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Listen normally on 4 lo [::1]:123 2026-03-10T12:29:44.977 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:5%2]:123 2026-03-10T12:29:44.977 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:44 ntpd[16107]: Listening on routing socket on fd #22 for interface updates 2026-03-10T12:29:45.939 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:45 ntpd[16092]: Soliciting pool server 144.76.43.40 2026-03-10T12:29:45.975 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:45 ntpd[16107]: Soliciting pool server 144.76.43.40 2026-03-10T12:29:46.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:46 ntpd[16092]: Soliciting pool server 157.90.24.29 2026-03-10T12:29:46.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:46 ntpd[16092]: Soliciting pool server 144.76.76.107 2026-03-10T12:29:46.974 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:46 ntpd[16107]: Soliciting pool server 157.90.24.29 2026-03-10T12:29:46.975 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:46 ntpd[16107]: Soliciting pool server 144.76.76.107 2026-03-10T12:29:47.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:47 ntpd[16092]: Soliciting pool server 194.36.144.87 2026-03-10T12:29:47.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:47 ntpd[16092]: Soliciting pool server 18.192.244.117 2026-03-10T12:29:47.939 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:47 ntpd[16092]: Soliciting pool server 62.128.1.18 2026-03-10T12:29:47.974 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:47 ntpd[16107]: Soliciting pool server 194.36.144.87 2026-03-10T12:29:47.974 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:47 ntpd[16107]: Soliciting pool server 18.192.244.117 2026-03-10T12:29:47.975 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:47 ntpd[16107]: Soliciting pool server 62.128.1.18 2026-03-10T12:29:48.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:48 ntpd[16092]: Soliciting pool server 188.245.170.46 2026-03-10T12:29:48.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:48 ntpd[16092]: Soliciting pool server 93.177.65.20 2026-03-10T12:29:48.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:48 ntpd[16092]: Soliciting pool server 185.248.189.10 2026-03-10T12:29:48.939 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:48 ntpd[16092]: Soliciting pool server 217.144.138.234 2026-03-10T12:29:48.973 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:48 ntpd[16107]: Soliciting pool server 188.245.170.46 2026-03-10T12:29:48.974 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:48 ntpd[16107]: Soliciting pool server 93.177.65.20 2026-03-10T12:29:48.974 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:48 ntpd[16107]: Soliciting pool server 185.248.189.10 2026-03-10T12:29:48.974 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:48 ntpd[16107]: Soliciting pool server 217.144.138.234 2026-03-10T12:29:49.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:49 ntpd[16092]: Soliciting pool server 82.165.178.31 2026-03-10T12:29:49.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:49 ntpd[16092]: Soliciting pool server 148.251.5.46 2026-03-10T12:29:49.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:49 ntpd[16092]: Soliciting pool server 162.159.200.1 2026-03-10T12:29:49.939 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:49 ntpd[16092]: Soliciting pool server 185.125.190.56 2026-03-10T12:29:49.973 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:49 ntpd[16107]: Soliciting pool server 82.165.178.31 2026-03-10T12:29:49.973 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:49 ntpd[16107]: Soliciting pool server 148.251.5.46 2026-03-10T12:29:49.973 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:49 ntpd[16107]: Soliciting pool server 162.159.200.1 2026-03-10T12:29:49.974 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:49 ntpd[16107]: Soliciting pool server 185.125.190.56 2026-03-10T12:29:50.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:50 ntpd[16092]: Soliciting pool server 185.125.190.58 2026-03-10T12:29:50.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:50 ntpd[16092]: Soliciting pool server 85.215.166.214 2026-03-10T12:29:50.938 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:50 ntpd[16092]: Soliciting pool server 162.159.200.123 2026-03-10T12:29:50.972 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:50 ntpd[16107]: Soliciting pool server 185.125.190.58 2026-03-10T12:29:50.973 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:50 ntpd[16107]: Soliciting pool server 85.215.166.214 2026-03-10T12:29:50.973 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:50 ntpd[16107]: Soliciting pool server 162.159.200.123 2026-03-10T12:29:52.963 INFO:teuthology.orchestra.run.vm01.stdout:10 Mar 12:29:52 ntpd[16092]: ntpd: time slew +0.000178 s 2026-03-10T12:29:52.963 INFO:teuthology.orchestra.run.vm01.stdout:ntpd: time slew +0.000178s 2026-03-10T12:29:52.983 INFO:teuthology.orchestra.run.vm01.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T12:29:52.983 INFO:teuthology.orchestra.run.vm01.stdout:============================================================================== 2026-03-10T12:29:52.983 INFO:teuthology.orchestra.run.vm01.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:52.983 INFO:teuthology.orchestra.run.vm01.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:52.983 INFO:teuthology.orchestra.run.vm01.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:52.983 INFO:teuthology.orchestra.run.vm01.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:52.983 INFO:teuthology.orchestra.run.vm01.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:53.995 INFO:teuthology.orchestra.run.vm05.stdout:10 Mar 12:29:53 ntpd[16107]: ntpd: time slew +0.008850 s 2026-03-10T12:29:53.995 INFO:teuthology.orchestra.run.vm05.stdout:ntpd: time slew +0.008850s 2026-03-10T12:29:54.014 INFO:teuthology.orchestra.run.vm05.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T12:29:54.014 INFO:teuthology.orchestra.run.vm05.stdout:============================================================================== 2026-03-10T12:29:54.014 INFO:teuthology.orchestra.run.vm05.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:54.014 INFO:teuthology.orchestra.run.vm05.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:54.014 INFO:teuthology.orchestra.run.vm05.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:54.014 INFO:teuthology.orchestra.run.vm05.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:54.014 INFO:teuthology.orchestra.run.vm05.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:29:54.015 INFO:teuthology.run_tasks:Running task install... 2026-03-10T12:29:54.017 DEBUG:teuthology.task.install:project ceph 2026-03-10T12:29:54.017 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-10T12:29:54.017 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-10T12:29:54.017 INFO:teuthology.task.install:Using flavor: default 2026-03-10T12:29:54.019 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-10T12:29:54.019 INFO:teuthology.task.install:extra packages: [] 2026-03-10T12:29:54.019 DEBUG:teuthology.orchestra.run.vm01:> sudo apt-key list | grep Ceph 2026-03-10T12:29:54.020 DEBUG:teuthology.orchestra.run.vm05:> sudo apt-key list | grep Ceph 2026-03-10T12:29:54.053 INFO:teuthology.orchestra.run.vm01.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-10T12:29:54.070 INFO:teuthology.orchestra.run.vm01.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-10T12:29:54.070 INFO:teuthology.orchestra.run.vm01.stdout:uid [ unknown] Ceph.com (release key) 2026-03-10T12:29:54.070 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-10T12:29:54.070 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-xmltodict, python3-jmespath on remote deb x86_64 2026-03-10T12:29:54.071 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:29:54.104 INFO:teuthology.orchestra.run.vm05.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-10T12:29:54.109 INFO:teuthology.orchestra.run.vm05.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-10T12:29:54.109 INFO:teuthology.orchestra.run.vm05.stdout:uid [ unknown] Ceph.com (release key) 2026-03-10T12:29:54.110 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-10T12:29:54.110 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-xmltodict, python3-jmespath on remote deb x86_64 2026-03-10T12:29:54.110 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:29:54.688 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/ 2026-03-10T12:29:54.689 INFO:teuthology.task.install.deb:Package version is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:29:54.720 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/ 2026-03-10T12:29:54.720 INFO:teuthology.task.install.deb:Package version is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:29:55.226 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:29:55.226 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-10T12:29:55.234 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:29:55.234 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-10T12:29:55.235 DEBUG:teuthology.orchestra.run.vm01:> sudo apt-get update 2026-03-10T12:29:55.243 DEBUG:teuthology.orchestra.run.vm05:> sudo apt-get update 2026-03-10T12:29:55.416 INFO:teuthology.orchestra.run.vm01.stdout:Hit:1 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T12:29:55.546 INFO:teuthology.orchestra.run.vm01.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T12:29:55.560 INFO:teuthology.orchestra.run.vm05.stdout:Hit:1 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T12:29:55.577 INFO:teuthology.orchestra.run.vm01.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T12:29:55.614 INFO:teuthology.orchestra.run.vm01.stdout:Hit:4 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T12:29:55.855 INFO:teuthology.orchestra.run.vm05.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T12:29:55.899 INFO:teuthology.orchestra.run.vm01.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy InRelease 2026-03-10T12:29:55.900 INFO:teuthology.orchestra.run.vm05.stdout:Ign:3 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy InRelease 2026-03-10T12:29:55.953 INFO:teuthology.orchestra.run.vm05.stdout:Hit:4 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T12:29:56.015 INFO:teuthology.orchestra.run.vm01.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release [7662 B] 2026-03-10T12:29:56.016 INFO:teuthology.orchestra.run.vm05.stdout:Get:5 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release [7662 B] 2026-03-10T12:29:56.050 INFO:teuthology.orchestra.run.vm05.stdout:Hit:6 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T12:29:56.132 INFO:teuthology.orchestra.run.vm01.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-10T12:29:56.132 INFO:teuthology.orchestra.run.vm05.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-10T12:29:56.252 INFO:teuthology.orchestra.run.vm01.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.1 kB] 2026-03-10T12:29:56.253 INFO:teuthology.orchestra.run.vm05.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.1 kB] 2026-03-10T12:29:56.324 INFO:teuthology.orchestra.run.vm01.stdout:Fetched 25.8 kB in 1s (27.7 kB/s) 2026-03-10T12:29:56.326 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 25.8 kB in 1s (27.7 kB/s) 2026-03-10T12:29:57.030 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:29:57.044 DEBUG:teuthology.orchestra.run.vm01:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=19.2.3-678-ge911bdeb-1jammy cephadm=19.2.3-678-ge911bdeb-1jammy ceph-mds=19.2.3-678-ge911bdeb-1jammy ceph-mgr=19.2.3-678-ge911bdeb-1jammy ceph-common=19.2.3-678-ge911bdeb-1jammy ceph-fuse=19.2.3-678-ge911bdeb-1jammy ceph-test=19.2.3-678-ge911bdeb-1jammy ceph-volume=19.2.3-678-ge911bdeb-1jammy radosgw=19.2.3-678-ge911bdeb-1jammy python3-rados=19.2.3-678-ge911bdeb-1jammy python3-rgw=19.2.3-678-ge911bdeb-1jammy python3-cephfs=19.2.3-678-ge911bdeb-1jammy python3-rbd=19.2.3-678-ge911bdeb-1jammy libcephfs2=19.2.3-678-ge911bdeb-1jammy libcephfs-dev=19.2.3-678-ge911bdeb-1jammy librados2=19.2.3-678-ge911bdeb-1jammy librbd1=19.2.3-678-ge911bdeb-1jammy rbd-fuse=19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:29:57.064 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:29:57.077 DEBUG:teuthology.orchestra.run.vm05:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=19.2.3-678-ge911bdeb-1jammy cephadm=19.2.3-678-ge911bdeb-1jammy ceph-mds=19.2.3-678-ge911bdeb-1jammy ceph-mgr=19.2.3-678-ge911bdeb-1jammy ceph-common=19.2.3-678-ge911bdeb-1jammy ceph-fuse=19.2.3-678-ge911bdeb-1jammy ceph-test=19.2.3-678-ge911bdeb-1jammy ceph-volume=19.2.3-678-ge911bdeb-1jammy radosgw=19.2.3-678-ge911bdeb-1jammy python3-rados=19.2.3-678-ge911bdeb-1jammy python3-rgw=19.2.3-678-ge911bdeb-1jammy python3-cephfs=19.2.3-678-ge911bdeb-1jammy python3-rbd=19.2.3-678-ge911bdeb-1jammy libcephfs2=19.2.3-678-ge911bdeb-1jammy libcephfs-dev=19.2.3-678-ge911bdeb-1jammy librados2=19.2.3-678-ge911bdeb-1jammy librbd1=19.2.3-678-ge911bdeb-1jammy rbd-fuse=19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:29:57.078 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:29:57.110 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:29:57.294 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:29:57.294 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:29:57.328 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:29:57.329 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout:The following additional packages will be installed: 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout: liboath0 libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-10T12:29:57.475 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsqlite3-mod-ceph 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: libthrift-0.16.0 lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-pytest python3-repoze.lru 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml python3-waitress python3-wcwidth python3-webob 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket python3-webtest python3-werkzeug python3-zc.lockfile 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: qttranslations5-l10n smartmontools socat unzip xmlstarlet zip 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout:Suggested packages: 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb readline-doc python3-beaker python-mako-doc 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python-natsort-doc httpd-wsgi libapache2-mod-python libapache2-mod-scgi 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: libjs-mochikit python-pecan-doc python-psutil-doc subversion 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python-pygments-doc ttf-bitstream-vera python-pyinotify-doc python3-dap 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python-sklearn-doc ipython3 python-waitress-doc python-webob-doc 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: python-webtest-doc python-werkzeug-doc python3-watchdog gsmartcontrol 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: smart-notifier mailx | mailutils 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout:Recommended packages: 2026-03-10T12:29:57.476 INFO:teuthology.orchestra.run.vm01.stdout: btrfs-tools 2026-03-10T12:29:57.498 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:29:57.499 INFO:teuthology.orchestra.run.vm05.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:29:57.499 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T12:29:57.499 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:29:57.499 INFO:teuthology.orchestra.run.vm05.stdout:The following additional packages will be installed: 2026-03-10T12:29:57.499 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-10T12:29:57.499 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsqlite3-mod-ceph 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: libthrift-0.16.0 lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T12:29:57.500 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-pytest python3-repoze.lru 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml python3-waitress python3-wcwidth python3-webob 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket python3-webtest python3-werkzeug python3-zc.lockfile 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: qttranslations5-l10n smartmontools socat unzip xmlstarlet zip 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout:Suggested packages: 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python3-influxdb readline-doc python3-beaker python-mako-doc 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python-natsort-doc httpd-wsgi libapache2-mod-python libapache2-mod-scgi 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: libjs-mochikit python-pecan-doc python-psutil-doc subversion 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python-pygments-doc ttf-bitstream-vera python-pyinotify-doc python3-dap 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python-sklearn-doc ipython3 python-waitress-doc python-webob-doc 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: python-webtest-doc python-werkzeug-doc python3-watchdog gsmartcontrol 2026-03-10T12:29:57.501 INFO:teuthology.orchestra.run.vm05.stdout: smart-notifier mailx | mailutils 2026-03-10T12:29:57.502 INFO:teuthology.orchestra.run.vm05.stdout:Recommended packages: 2026-03-10T12:29:57.502 INFO:teuthology.orchestra.run.vm05.stdout: btrfs-tools 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout:The following NEW packages will be installed: 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-dev libcephfs2 libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: liblua5.3-dev libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 libreadline-dev 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 libsqlite3-mod-ceph libthrift-0.16.0 lua-any lua-sec lua-socket 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: lua5.1 luarocks nvme-cli pkg-config python-asyncssh-doc 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:29:57.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-pytest python3-rados python3-rbd 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze.lru python3-requests-oauthlib python3-rgw python3-routes 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-threadpoolctl python3-toml python3-waitress python3-wcwidth 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse smartmontools 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: socat unzip xmlstarlet zip 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be upgraded: 2026-03-10T12:29:57.514 INFO:teuthology.orchestra.run.vm01.stdout: librados2 librbd1 2026-03-10T12:29:57.537 INFO:teuthology.orchestra.run.vm05.stdout:The following NEW packages will be installed: 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-dev libcephfs2 libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: liblua5.3-dev libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 libreadline-dev 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: librgw2 libsqlite3-mod-ceph libthrift-0.16.0 lua-any lua-sec lua-socket 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: lua5.1 luarocks nvme-cli pkg-config python-asyncssh-doc 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T12:29:57.538 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-pytest python3-rados python3-rbd 2026-03-10T12:29:57.539 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze.lru python3-requests-oauthlib python3-rgw python3-routes 2026-03-10T12:29:57.539 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T12:29:57.539 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T12:29:57.539 INFO:teuthology.orchestra.run.vm05.stdout: python3-threadpoolctl python3-toml python3-waitress python3-wcwidth 2026-03-10T12:29:57.539 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:29:57.539 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse smartmontools 2026-03-10T12:29:57.539 INFO:teuthology.orchestra.run.vm05.stdout: socat unzip xmlstarlet zip 2026-03-10T12:29:57.539 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be upgraded: 2026-03-10T12:29:57.540 INFO:teuthology.orchestra.run.vm05.stdout: librados2 librbd1 2026-03-10T12:29:58.016 INFO:teuthology.orchestra.run.vm05.stdout:2 upgraded, 107 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:29:58.016 INFO:teuthology.orchestra.run.vm05.stdout:Need to get 178 MB of archives. 2026-03-10T12:29:58.016 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 782 MB of additional disk space will be used. 2026-03-10T12:29:58.016 INFO:teuthology.orchestra.run.vm05.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-10T12:29:58.031 INFO:teuthology.orchestra.run.vm01.stdout:2 upgraded, 107 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:29:58.031 INFO:teuthology.orchestra.run.vm01.stdout:Need to get 178 MB of archives. 2026-03-10T12:29:58.031 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 782 MB of additional disk space will be used. 2026-03-10T12:29:58.031 INFO:teuthology.orchestra.run.vm01.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-10T12:29:58.104 INFO:teuthology.orchestra.run.vm05.stdout:Get:2 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 19.2.3-678-ge911bdeb-1jammy [3257 kB] 2026-03-10T12:29:58.113 INFO:teuthology.orchestra.run.vm01.stdout:Get:2 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 19.2.3-678-ge911bdeb-1jammy [3257 kB] 2026-03-10T12:29:58.478 INFO:teuthology.orchestra.run.vm05.stdout:Get:3 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-10T12:29:58.493 INFO:teuthology.orchestra.run.vm05.stdout:Get:4 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-10T12:29:58.505 INFO:teuthology.orchestra.run.vm01.stdout:Get:3 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-10T12:29:58.520 INFO:teuthology.orchestra.run.vm01.stdout:Get:4 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-10T12:29:58.588 INFO:teuthology.orchestra.run.vm05.stdout:Get:5 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-10T12:29:58.617 INFO:teuthology.orchestra.run.vm01.stdout:Get:5 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-10T12:29:58.875 INFO:teuthology.orchestra.run.vm05.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 19.2.3-678-ge911bdeb-1jammy [3597 kB] 2026-03-10T12:29:58.919 INFO:teuthology.orchestra.run.vm05.stdout:Get:7 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-10T12:29:59.037 INFO:teuthology.orchestra.run.vm05.stdout:Get:8 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-10T12:29:59.042 INFO:teuthology.orchestra.run.vm05.stdout:Get:9 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-10T12:29:59.044 INFO:teuthology.orchestra.run.vm05.stdout:Get:10 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-10T12:29:59.045 INFO:teuthology.orchestra.run.vm05.stdout:Get:11 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-10T12:29:59.045 INFO:teuthology.orchestra.run.vm05.stdout:Get:12 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-10T12:29:59.045 INFO:teuthology.orchestra.run.vm05.stdout:Get:13 https://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-10T12:29:59.060 INFO:teuthology.orchestra.run.vm01.stdout:Get:6 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-10T12:29:59.060 INFO:teuthology.orchestra.run.vm01.stdout:Get:7 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-10T12:29:59.096 INFO:teuthology.orchestra.run.vm05.stdout:Get:14 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 19.2.3-678-ge911bdeb-1jammy [979 kB] 2026-03-10T12:29:59.099 INFO:teuthology.orchestra.run.vm05.stdout:Get:15 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 19.2.3-678-ge911bdeb-1jammy [357 kB] 2026-03-10T12:29:59.100 INFO:teuthology.orchestra.run.vm05.stdout:Get:16 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 19.2.3-678-ge911bdeb-1jammy [32.9 kB] 2026-03-10T12:29:59.100 INFO:teuthology.orchestra.run.vm05.stdout:Get:17 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 19.2.3-678-ge911bdeb-1jammy [184 kB] 2026-03-10T12:29:59.101 INFO:teuthology.orchestra.run.vm05.stdout:Get:18 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 19.2.3-678-ge911bdeb-1jammy [70.1 kB] 2026-03-10T12:29:59.101 INFO:teuthology.orchestra.run.vm05.stdout:Get:19 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 19.2.3-678-ge911bdeb-1jammy [334 kB] 2026-03-10T12:29:59.135 INFO:teuthology.orchestra.run.vm05.stdout:Get:20 https://archive.ubuntu.com/ubuntu jammy/main amd64 libreadline-dev amd64 8.1.2-1 [166 kB] 2026-03-10T12:29:59.136 INFO:teuthology.orchestra.run.vm05.stdout:Get:21 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblua5.3-dev amd64 5.3.6-1build1 [167 kB] 2026-03-10T12:29:59.137 INFO:teuthology.orchestra.run.vm05.stdout:Get:22 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua5.1 amd64 5.1.5-8.1build4 [94.6 kB] 2026-03-10T12:29:59.138 INFO:teuthology.orchestra.run.vm05.stdout:Get:23 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-any all 27ubuntu1 [5034 B] 2026-03-10T12:29:59.160 INFO:teuthology.orchestra.run.vm01.stdout:Get:8 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-10T12:29:59.161 INFO:teuthology.orchestra.run.vm01.stdout:Get:9 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-10T12:29:59.161 INFO:teuthology.orchestra.run.vm01.stdout:Get:10 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-10T12:29:59.161 INFO:teuthology.orchestra.run.vm01.stdout:Get:11 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-10T12:29:59.162 INFO:teuthology.orchestra.run.vm01.stdout:Get:12 https://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-10T12:29:59.164 INFO:teuthology.orchestra.run.vm01.stdout:Get:13 https://archive.ubuntu.com/ubuntu jammy/main amd64 libreadline-dev amd64 8.1.2-1 [166 kB] 2026-03-10T12:29:59.165 INFO:teuthology.orchestra.run.vm01.stdout:Get:14 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblua5.3-dev amd64 5.3.6-1build1 [167 kB] 2026-03-10T12:29:59.166 INFO:teuthology.orchestra.run.vm01.stdout:Get:15 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua5.1 amd64 5.1.5-8.1build4 [94.6 kB] 2026-03-10T12:29:59.201 INFO:teuthology.orchestra.run.vm05.stdout:Get:24 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 19.2.3-678-ge911bdeb-1jammy [6935 kB] 2026-03-10T12:29:59.234 INFO:teuthology.orchestra.run.vm05.stdout:Get:25 https://archive.ubuntu.com/ubuntu jammy/main amd64 zip amd64 3.0-12build2 [176 kB] 2026-03-10T12:29:59.237 INFO:teuthology.orchestra.run.vm05.stdout:Get:26 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 unzip amd64 6.0-26ubuntu3.2 [175 kB] 2026-03-10T12:29:59.241 INFO:teuthology.orchestra.run.vm05.stdout:Get:27 https://archive.ubuntu.com/ubuntu jammy/universe amd64 luarocks all 3.8.0+dfsg1-1 [140 kB] 2026-03-10T12:29:59.244 INFO:teuthology.orchestra.run.vm05.stdout:Get:28 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-10T12:29:59.245 INFO:teuthology.orchestra.run.vm05.stdout:Get:29 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-10T12:29:59.245 INFO:teuthology.orchestra.run.vm05.stdout:Get:30 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-10T12:29:59.263 INFO:teuthology.orchestra.run.vm01.stdout:Get:16 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-any all 27ubuntu1 [5034 B] 2026-03-10T12:29:59.263 INFO:teuthology.orchestra.run.vm01.stdout:Get:17 https://archive.ubuntu.com/ubuntu jammy/main amd64 zip amd64 3.0-12build2 [176 kB] 2026-03-10T12:29:59.268 INFO:teuthology.orchestra.run.vm01.stdout:Get:18 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 unzip amd64 6.0-26ubuntu3.2 [175 kB] 2026-03-10T12:29:59.310 INFO:teuthology.orchestra.run.vm01.stdout:Get:19 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 19.2.3-678-ge911bdeb-1jammy [3597 kB] 2026-03-10T12:29:59.338 INFO:teuthology.orchestra.run.vm05.stdout:Get:31 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-10T12:29:59.364 INFO:teuthology.orchestra.run.vm01.stdout:Get:20 https://archive.ubuntu.com/ubuntu jammy/universe amd64 luarocks all 3.8.0+dfsg1-1 [140 kB] 2026-03-10T12:29:59.367 INFO:teuthology.orchestra.run.vm01.stdout:Get:21 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-10T12:29:59.368 INFO:teuthology.orchestra.run.vm01.stdout:Get:22 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-10T12:29:59.368 INFO:teuthology.orchestra.run.vm01.stdout:Get:23 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-10T12:29:59.371 INFO:teuthology.orchestra.run.vm01.stdout:Get:24 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-10T12:29:59.371 INFO:teuthology.orchestra.run.vm01.stdout:Get:25 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-10T12:29:59.371 INFO:teuthology.orchestra.run.vm01.stdout:Get:26 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-10T12:29:59.465 INFO:teuthology.orchestra.run.vm05.stdout:Get:32 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-10T12:29:59.465 INFO:teuthology.orchestra.run.vm01.stdout:Get:27 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-10T12:29:59.465 INFO:teuthology.orchestra.run.vm05.stdout:Get:33 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-10T12:29:59.465 INFO:teuthology.orchestra.run.vm01.stdout:Get:28 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-10T12:29:59.466 INFO:teuthology.orchestra.run.vm01.stdout:Get:29 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-10T12:29:59.466 INFO:teuthology.orchestra.run.vm05.stdout:Get:34 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-10T12:29:59.466 INFO:teuthology.orchestra.run.vm05.stdout:Get:35 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-10T12:29:59.466 INFO:teuthology.orchestra.run.vm05.stdout:Get:36 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-10T12:29:59.466 INFO:teuthology.orchestra.run.vm05.stdout:Get:37 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-10T12:29:59.467 INFO:teuthology.orchestra.run.vm05.stdout:Get:38 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-10T12:29:59.468 INFO:teuthology.orchestra.run.vm05.stdout:Get:39 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-logutils all 0.3.3-8 [17.6 kB] 2026-03-10T12:29:59.468 INFO:teuthology.orchestra.run.vm05.stdout:Get:40 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-mako all 1.1.3+ds1-2ubuntu0.1 [60.5 kB] 2026-03-10T12:29:59.534 INFO:teuthology.orchestra.run.vm05.stdout:Get:41 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplegeneric all 0.8.1-3 [11.3 kB] 2026-03-10T12:29:59.566 INFO:teuthology.orchestra.run.vm01.stdout:Get:30 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-10T12:29:59.567 INFO:teuthology.orchestra.run.vm05.stdout:Get:42 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-singledispatch all 3.4.0.3-3 [7320 B] 2026-03-10T12:29:59.568 INFO:teuthology.orchestra.run.vm05.stdout:Get:43 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-10T12:29:59.569 INFO:teuthology.orchestra.run.vm05.stdout:Get:44 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-waitress all 1.4.4-1.1ubuntu1.1 [47.0 kB] 2026-03-10T12:29:59.570 INFO:teuthology.orchestra.run.vm05.stdout:Get:45 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempita all 0.5.2-6ubuntu1 [15.1 kB] 2026-03-10T12:29:59.571 INFO:teuthology.orchestra.run.vm05.stdout:Get:46 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-paste all 3.5.0+dfsg1-1 [456 kB] 2026-03-10T12:29:59.573 INFO:teuthology.orchestra.run.vm01.stdout:Get:31 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-10T12:29:59.573 INFO:teuthology.orchestra.run.vm01.stdout:Get:32 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-logutils all 0.3.3-8 [17.6 kB] 2026-03-10T12:29:59.573 INFO:teuthology.orchestra.run.vm01.stdout:Get:33 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-mako all 1.1.3+ds1-2ubuntu0.1 [60.5 kB] 2026-03-10T12:29:59.576 INFO:teuthology.orchestra.run.vm01.stdout:Get:34 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplegeneric all 0.8.1-3 [11.3 kB] 2026-03-10T12:29:59.576 INFO:teuthology.orchestra.run.vm01.stdout:Get:35 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-singledispatch all 3.4.0.3-3 [7320 B] 2026-03-10T12:29:59.577 INFO:teuthology.orchestra.run.vm01.stdout:Get:36 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-10T12:29:59.582 INFO:teuthology.orchestra.run.vm05.stdout:Get:47 https://archive.ubuntu.com/ubuntu jammy/main amd64 python-pastedeploy-tpl all 2.1.1-1 [4892 B] 2026-03-10T12:29:59.582 INFO:teuthology.orchestra.run.vm05.stdout:Get:48 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastedeploy all 2.1.1-1 [26.6 kB] 2026-03-10T12:29:59.606 INFO:teuthology.orchestra.run.vm05.stdout:Get:49 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 19.2.3-678-ge911bdeb-1jammy [112 kB] 2026-03-10T12:29:59.606 INFO:teuthology.orchestra.run.vm05.stdout:Get:50 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 19.2.3-678-ge911bdeb-1jammy [470 kB] 2026-03-10T12:29:59.632 INFO:teuthology.orchestra.run.vm05.stdout:Get:51 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-webtest all 2.0.35-1 [28.5 kB] 2026-03-10T12:29:59.633 INFO:teuthology.orchestra.run.vm05.stdout:Get:52 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pecan all 1.3.3-4ubuntu2 [87.3 kB] 2026-03-10T12:29:59.667 INFO:teuthology.orchestra.run.vm01.stdout:Get:37 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-waitress all 1.4.4-1.1ubuntu1.1 [47.0 kB] 2026-03-10T12:29:59.668 INFO:teuthology.orchestra.run.vm01.stdout:Get:38 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempita all 0.5.2-6ubuntu1 [15.1 kB] 2026-03-10T12:29:59.668 INFO:teuthology.orchestra.run.vm01.stdout:Get:39 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-paste all 3.5.0+dfsg1-1 [456 kB] 2026-03-10T12:29:59.681 INFO:teuthology.orchestra.run.vm05.stdout:Get:53 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-werkzeug all 2.0.2+dfsg1-1ubuntu0.22.04.3 [181 kB] 2026-03-10T12:29:59.684 INFO:teuthology.orchestra.run.vm05.stdout:Get:54 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-10T12:29:59.686 INFO:teuthology.orchestra.run.vm05.stdout:Get:55 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-10T12:29:59.690 INFO:teuthology.orchestra.run.vm05.stdout:Get:56 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-10T12:29:59.691 INFO:teuthology.orchestra.run.vm05.stdout:Get:57 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-10T12:29:59.693 INFO:teuthology.orchestra.run.vm05.stdout:Get:58 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-10T12:29:59.768 INFO:teuthology.orchestra.run.vm01.stdout:Get:40 https://archive.ubuntu.com/ubuntu jammy/main amd64 python-pastedeploy-tpl all 2.1.1-1 [4892 B] 2026-03-10T12:29:59.769 INFO:teuthology.orchestra.run.vm01.stdout:Get:41 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastedeploy all 2.1.1-1 [26.6 kB] 2026-03-10T12:29:59.769 INFO:teuthology.orchestra.run.vm01.stdout:Get:42 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-webtest all 2.0.35-1 [28.5 kB] 2026-03-10T12:29:59.769 INFO:teuthology.orchestra.run.vm01.stdout:Get:43 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pecan all 1.3.3-4ubuntu2 [87.3 kB] 2026-03-10T12:29:59.772 INFO:teuthology.orchestra.run.vm01.stdout:Get:44 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-werkzeug all 2.0.2+dfsg1-1ubuntu0.22.04.3 [181 kB] 2026-03-10T12:29:59.777 INFO:teuthology.orchestra.run.vm01.stdout:Get:45 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-10T12:29:59.781 INFO:teuthology.orchestra.run.vm01.stdout:Get:46 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-10T12:29:59.818 INFO:teuthology.orchestra.run.vm05.stdout:Get:59 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 19.2.3-678-ge911bdeb-1jammy [26.5 MB] 2026-03-10T12:29:59.870 INFO:teuthology.orchestra.run.vm01.stdout:Get:47 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-10T12:29:59.870 INFO:teuthology.orchestra.run.vm01.stdout:Get:48 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-10T12:29:59.873 INFO:teuthology.orchestra.run.vm01.stdout:Get:49 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-10T12:29:59.873 INFO:teuthology.orchestra.run.vm05.stdout:Get:60 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-10T12:29:59.937 INFO:teuthology.orchestra.run.vm05.stdout:Get:61 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-10T12:29:59.938 INFO:teuthology.orchestra.run.vm05.stdout:Get:62 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-10T12:29:59.972 INFO:teuthology.orchestra.run.vm05.stdout:Get:63 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-10T12:29:59.973 INFO:teuthology.orchestra.run.vm05.stdout:Get:64 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-10T12:29:59.973 INFO:teuthology.orchestra.run.vm05.stdout:Get:65 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-10T12:30:00.035 INFO:teuthology.orchestra.run.vm05.stdout:Get:66 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-10T12:30:00.035 INFO:teuthology.orchestra.run.vm05.stdout:Get:67 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-10T12:30:00.035 INFO:teuthology.orchestra.run.vm05.stdout:Get:68 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-10T12:30:00.039 INFO:teuthology.orchestra.run.vm01.stdout:Get:50 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-10T12:30:00.040 INFO:teuthology.orchestra.run.vm01.stdout:Get:51 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-10T12:30:00.040 INFO:teuthology.orchestra.run.vm01.stdout:Get:52 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-10T12:30:00.062 INFO:teuthology.orchestra.run.vm01.stdout:Get:53 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 19.2.3-678-ge911bdeb-1jammy [979 kB] 2026-03-10T12:30:00.139 INFO:teuthology.orchestra.run.vm01.stdout:Get:54 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-10T12:30:00.139 INFO:teuthology.orchestra.run.vm01.stdout:Get:55 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-10T12:30:00.140 INFO:teuthology.orchestra.run.vm01.stdout:Get:56 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-10T12:30:00.141 INFO:teuthology.orchestra.run.vm01.stdout:Get:57 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-10T12:30:00.141 INFO:teuthology.orchestra.run.vm01.stdout:Get:58 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-10T12:30:00.142 INFO:teuthology.orchestra.run.vm01.stdout:Get:59 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-10T12:30:00.173 INFO:teuthology.orchestra.run.vm05.stdout:Get:69 https://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-10T12:30:00.174 INFO:teuthology.orchestra.run.vm05.stdout:Get:70 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-10T12:30:00.175 INFO:teuthology.orchestra.run.vm05.stdout:Get:71 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-10T12:30:00.175 INFO:teuthology.orchestra.run.vm05.stdout:Get:72 https://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-10T12:30:00.177 INFO:teuthology.orchestra.run.vm05.stdout:Get:73 https://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-10T12:30:00.272 INFO:teuthology.orchestra.run.vm05.stdout:Get:74 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-socket amd64 3.0~rc1+git+ac3201d-6 [78.9 kB] 2026-03-10T12:30:00.272 INFO:teuthology.orchestra.run.vm05.stdout:Get:75 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-sec amd64 1.0.2-1 [37.6 kB] 2026-03-10T12:30:00.272 INFO:teuthology.orchestra.run.vm05.stdout:Get:76 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-10T12:30:00.281 INFO:teuthology.orchestra.run.vm01.stdout:Get:60 https://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-10T12:30:00.282 INFO:teuthology.orchestra.run.vm01.stdout:Get:61 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-10T12:30:00.283 INFO:teuthology.orchestra.run.vm01.stdout:Get:62 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-10T12:30:00.309 INFO:teuthology.orchestra.run.vm01.stdout:Get:63 https://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-10T12:30:00.370 INFO:teuthology.orchestra.run.vm05.stdout:Get:77 https://archive.ubuntu.com/ubuntu jammy/main amd64 pkg-config amd64 0.29.2-1ubuntu3 [48.2 kB] 2026-03-10T12:30:00.370 INFO:teuthology.orchestra.run.vm05.stdout:Get:78 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-10T12:30:00.372 INFO:teuthology.orchestra.run.vm05.stdout:Get:79 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-10T12:30:00.372 INFO:teuthology.orchestra.run.vm05.stdout:Get:80 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastescript all 2.0.2-4 [54.6 kB] 2026-03-10T12:30:00.372 INFO:teuthology.orchestra.run.vm05.stdout:Get:81 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-10T12:30:00.373 INFO:teuthology.orchestra.run.vm05.stdout:Get:82 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-10T12:30:00.428 INFO:teuthology.orchestra.run.vm05.stdout:Get:83 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-10T12:30:00.430 INFO:teuthology.orchestra.run.vm05.stdout:Get:84 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-10T12:30:00.444 INFO:teuthology.orchestra.run.vm05.stdout:Get:85 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pyinotify all 0.9.6-1.3 [24.8 kB] 2026-03-10T12:30:00.446 INFO:teuthology.orchestra.run.vm01.stdout:Get:64 https://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-10T12:30:00.447 INFO:teuthology.orchestra.run.vm01.stdout:Get:65 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-socket amd64 3.0~rc1+git+ac3201d-6 [78.9 kB] 2026-03-10T12:30:00.448 INFO:teuthology.orchestra.run.vm01.stdout:Get:66 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-sec amd64 1.0.2-1 [37.6 kB] 2026-03-10T12:30:00.448 INFO:teuthology.orchestra.run.vm01.stdout:Get:67 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-10T12:30:00.451 INFO:teuthology.orchestra.run.vm01.stdout:Get:68 https://archive.ubuntu.com/ubuntu jammy/main amd64 pkg-config amd64 0.29.2-1ubuntu3 [48.2 kB] 2026-03-10T12:30:00.451 INFO:teuthology.orchestra.run.vm01.stdout:Get:69 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-10T12:30:00.527 INFO:teuthology.orchestra.run.vm05.stdout:Get:86 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-10T12:30:00.527 INFO:teuthology.orchestra.run.vm05.stdout:Get:87 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-10T12:30:00.530 INFO:teuthology.orchestra.run.vm05.stdout:Get:88 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-10T12:30:00.530 INFO:teuthology.orchestra.run.vm05.stdout:Get:89 https://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-10T12:30:00.554 INFO:teuthology.orchestra.run.vm01.stdout:Get:70 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-10T12:30:00.554 INFO:teuthology.orchestra.run.vm01.stdout:Get:71 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastescript all 2.0.2-4 [54.6 kB] 2026-03-10T12:30:00.554 INFO:teuthology.orchestra.run.vm01.stdout:Get:72 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-10T12:30:00.584 INFO:teuthology.orchestra.run.vm01.stdout:Get:73 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 19.2.3-678-ge911bdeb-1jammy [357 kB] 2026-03-10T12:30:00.585 INFO:teuthology.orchestra.run.vm01.stdout:Get:74 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 19.2.3-678-ge911bdeb-1jammy [32.9 kB] 2026-03-10T12:30:00.587 INFO:teuthology.orchestra.run.vm01.stdout:Get:75 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 19.2.3-678-ge911bdeb-1jammy [184 kB] 2026-03-10T12:30:00.608 INFO:teuthology.orchestra.run.vm01.stdout:Get:76 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 19.2.3-678-ge911bdeb-1jammy [70.1 kB] 2026-03-10T12:30:00.686 INFO:teuthology.orchestra.run.vm01.stdout:Get:77 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-10T12:30:00.687 INFO:teuthology.orchestra.run.vm01.stdout:Get:78 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-10T12:30:00.688 INFO:teuthology.orchestra.run.vm01.stdout:Get:79 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-10T12:30:00.691 INFO:teuthology.orchestra.run.vm01.stdout:Get:80 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pyinotify all 0.9.6-1.3 [24.8 kB] 2026-03-10T12:30:00.691 INFO:teuthology.orchestra.run.vm01.stdout:Get:81 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-10T12:30:00.692 INFO:teuthology.orchestra.run.vm01.stdout:Get:82 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-10T12:30:00.702 INFO:teuthology.orchestra.run.vm01.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 19.2.3-678-ge911bdeb-1jammy [334 kB] 2026-03-10T12:30:00.708 INFO:teuthology.orchestra.run.vm05.stdout:Get:90 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-10T12:30:00.757 INFO:teuthology.orchestra.run.vm01.stdout:Get:84 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-10T12:30:00.758 INFO:teuthology.orchestra.run.vm01.stdout:Get:85 https://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-10T12:30:00.807 INFO:teuthology.orchestra.run.vm01.stdout:Get:86 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-10T12:30:01.033 INFO:teuthology.orchestra.run.vm01.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 19.2.3-678-ge911bdeb-1jammy [6935 kB] 2026-03-10T12:30:03.109 INFO:teuthology.orchestra.run.vm05.stdout:Get:91 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 19.2.3-678-ge911bdeb-1jammy [5178 kB] 2026-03-10T12:30:03.754 INFO:teuthology.orchestra.run.vm05.stdout:Get:92 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 19.2.3-678-ge911bdeb-1jammy [248 kB] 2026-03-10T12:30:03.763 INFO:teuthology.orchestra.run.vm05.stdout:Get:93 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 19.2.3-678-ge911bdeb-1jammy [125 kB] 2026-03-10T12:30:03.765 INFO:teuthology.orchestra.run.vm05.stdout:Get:94 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 19.2.3-678-ge911bdeb-1jammy [1081 kB] 2026-03-10T12:30:03.954 INFO:teuthology.orchestra.run.vm05.stdout:Get:95 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 19.2.3-678-ge911bdeb-1jammy [6239 kB] 2026-03-10T12:30:04.635 INFO:teuthology.orchestra.run.vm01.stdout:Get:88 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 19.2.3-678-ge911bdeb-1jammy [112 kB] 2026-03-10T12:30:04.648 INFO:teuthology.orchestra.run.vm01.stdout:Get:89 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 19.2.3-678-ge911bdeb-1jammy [470 kB] 2026-03-10T12:30:04.713 INFO:teuthology.orchestra.run.vm05.stdout:Get:96 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 19.2.3-678-ge911bdeb-1jammy [23.0 MB] 2026-03-10T12:30:04.881 INFO:teuthology.orchestra.run.vm01.stdout:Get:90 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 19.2.3-678-ge911bdeb-1jammy [26.5 MB] 2026-03-10T12:30:08.846 INFO:teuthology.orchestra.run.vm05.stdout:Get:97 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 19.2.3-678-ge911bdeb-1jammy [14.2 kB] 2026-03-10T12:30:08.846 INFO:teuthology.orchestra.run.vm05.stdout:Get:98 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 19.2.3-678-ge911bdeb-1jammy [1173 kB] 2026-03-10T12:30:09.070 INFO:teuthology.orchestra.run.vm05.stdout:Get:99 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 19.2.3-678-ge911bdeb-1jammy [2503 kB] 2026-03-10T12:30:09.598 INFO:teuthology.orchestra.run.vm05.stdout:Get:100 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 19.2.3-678-ge911bdeb-1jammy [798 kB] 2026-03-10T12:30:09.719 INFO:teuthology.orchestra.run.vm05.stdout:Get:101 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 19.2.3-678-ge911bdeb-1jammy [157 kB] 2026-03-10T12:30:09.792 INFO:teuthology.orchestra.run.vm05.stdout:Get:102 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 19.2.3-678-ge911bdeb-1jammy [2396 kB] 2026-03-10T12:30:10.255 INFO:teuthology.orchestra.run.vm05.stdout:Get:103 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 19.2.3-678-ge911bdeb-1jammy [8625 kB] 2026-03-10T12:30:11.982 INFO:teuthology.orchestra.run.vm05.stdout:Get:104 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 19.2.3-678-ge911bdeb-1jammy [14.3 kB] 2026-03-10T12:30:12.008 INFO:teuthology.orchestra.run.vm05.stdout:Get:105 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 19.2.3-678-ge911bdeb-1jammy [52.1 MB] 2026-03-10T12:30:14.296 INFO:teuthology.orchestra.run.vm01.stdout:Get:91 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 19.2.3-678-ge911bdeb-1jammy [5178 kB] 2026-03-10T12:30:15.110 INFO:teuthology.orchestra.run.vm01.stdout:Get:92 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 19.2.3-678-ge911bdeb-1jammy [248 kB] 2026-03-10T12:30:15.124 INFO:teuthology.orchestra.run.vm01.stdout:Get:93 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 19.2.3-678-ge911bdeb-1jammy [125 kB] 2026-03-10T12:30:15.128 INFO:teuthology.orchestra.run.vm01.stdout:Get:94 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 19.2.3-678-ge911bdeb-1jammy [1081 kB] 2026-03-10T12:30:15.329 INFO:teuthology.orchestra.run.vm01.stdout:Get:95 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 19.2.3-678-ge911bdeb-1jammy [6239 kB] 2026-03-10T12:30:16.136 INFO:teuthology.orchestra.run.vm01.stdout:Get:96 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 19.2.3-678-ge911bdeb-1jammy [23.0 MB] 2026-03-10T12:30:18.237 INFO:teuthology.orchestra.run.vm01.stdout:Get:97 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 19.2.3-678-ge911bdeb-1jammy [14.2 kB] 2026-03-10T12:30:18.237 INFO:teuthology.orchestra.run.vm01.stdout:Get:98 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 19.2.3-678-ge911bdeb-1jammy [1173 kB] 2026-03-10T12:30:18.346 INFO:teuthology.orchestra.run.vm01.stdout:Get:99 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 19.2.3-678-ge911bdeb-1jammy [2503 kB] 2026-03-10T12:30:18.553 INFO:teuthology.orchestra.run.vm01.stdout:Get:100 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 19.2.3-678-ge911bdeb-1jammy [798 kB] 2026-03-10T12:30:18.579 INFO:teuthology.orchestra.run.vm01.stdout:Get:101 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 19.2.3-678-ge911bdeb-1jammy [157 kB] 2026-03-10T12:30:18.581 INFO:teuthology.orchestra.run.vm01.stdout:Get:102 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 19.2.3-678-ge911bdeb-1jammy [2396 kB] 2026-03-10T12:30:18.785 INFO:teuthology.orchestra.run.vm01.stdout:Get:103 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 19.2.3-678-ge911bdeb-1jammy [8625 kB] 2026-03-10T12:30:19.358 INFO:teuthology.orchestra.run.vm01.stdout:Get:104 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 19.2.3-678-ge911bdeb-1jammy [14.3 kB] 2026-03-10T12:30:19.358 INFO:teuthology.orchestra.run.vm01.stdout:Get:105 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 19.2.3-678-ge911bdeb-1jammy [52.1 MB] 2026-03-10T12:30:20.625 INFO:teuthology.orchestra.run.vm05.stdout:Get:106 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 19.2.3-678-ge911bdeb-1jammy [135 kB] 2026-03-10T12:30:20.625 INFO:teuthology.orchestra.run.vm05.stdout:Get:107 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 19.2.3-678-ge911bdeb-1jammy [41.0 kB] 2026-03-10T12:30:20.640 INFO:teuthology.orchestra.run.vm05.stdout:Get:108 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 19.2.3-678-ge911bdeb-1jammy [13.7 MB] 2026-03-10T12:30:21.904 INFO:teuthology.orchestra.run.vm01.stdout:Get:106 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 19.2.3-678-ge911bdeb-1jammy [135 kB] 2026-03-10T12:30:21.905 INFO:teuthology.orchestra.run.vm01.stdout:Get:107 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 19.2.3-678-ge911bdeb-1jammy [41.0 kB] 2026-03-10T12:30:21.905 INFO:teuthology.orchestra.run.vm01.stdout:Get:108 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 19.2.3-678-ge911bdeb-1jammy [13.7 MB] 2026-03-10T12:30:22.030 INFO:teuthology.orchestra.run.vm05.stdout:Get:109 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 19.2.3-678-ge911bdeb-1jammy [92.2 kB] 2026-03-10T12:30:22.347 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 178 MB in 24s (7272 kB/s) 2026-03-10T12:30:22.400 INFO:teuthology.orchestra.run.vm01.stdout:Get:109 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 19.2.3-678-ge911bdeb-1jammy [92.2 kB] 2026-03-10T12:30:22.597 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-10T12:30:22.629 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 111717 files and directories currently installed.) 2026-03-10T12:30:22.631 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../000-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-10T12:30:22.740 INFO:teuthology.orchestra.run.vm01.stdout:Fetched 178 MB in 25s (7149 kB/s) 2026-03-10T12:30:22.909 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T12:30:23.027 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-10T12:30:23.060 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 111717 files and directories currently installed.) 2026-03-10T12:30:23.062 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../000-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-10T12:30:23.153 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T12:30:23.166 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-10T12:30:23.170 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../001-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-10T12:30:23.171 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T12:30:23.176 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-10T12:30:23.182 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../001-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-10T12:30:23.182 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T12:30:23.189 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-10T12:30:23.194 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../002-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-10T12:30:23.196 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T12:30:23.198 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-10T12:30:23.204 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../002-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-10T12:30:23.205 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T12:30:23.219 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-10T12:30:23.225 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../003-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T12:30:23.227 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-10T12:30:23.229 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:23.234 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../003-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T12:30:23.239 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:23.280 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-10T12:30:23.281 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-10T12:30:23.286 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../004-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T12:30:23.287 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../004-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T12:30:23.287 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:23.288 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:23.308 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-10T12:30:23.309 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-10T12:30:23.314 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../005-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T12:30:23.315 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:23.315 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../005-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T12:30:23.316 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:23.348 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-10T12:30:23.354 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../006-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-10T12:30:23.387 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-10T12:30:23.387 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T12:30:23.393 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../006-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-10T12:30:23.394 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T12:30:23.412 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../007-librbd1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.414 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librbd1 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T12:30:23.423 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../007-librbd1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.426 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking librbd1 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T12:30:23.532 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../008-librados2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.535 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librados2 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T12:30:23.536 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../008-librados2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.540 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking librados2 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T12:30:23.624 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libnbd0. 2026-03-10T12:30:23.627 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libnbd0. 2026-03-10T12:30:23.631 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../009-libnbd0_1.10.5-1_amd64.deb ... 2026-03-10T12:30:23.632 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-10T12:30:23.634 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../009-libnbd0_1.10.5-1_amd64.deb ... 2026-03-10T12:30:23.635 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-10T12:30:23.649 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libcephfs2. 2026-03-10T12:30:23.654 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libcephfs2. 2026-03-10T12:30:23.655 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../010-libcephfs2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.656 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.660 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../010-libcephfs2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.660 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.685 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rados. 2026-03-10T12:30:23.688 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-rados. 2026-03-10T12:30:23.692 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../011-python3-rados_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.692 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.694 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../011-python3-rados_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.695 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.738 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-10T12:30:23.739 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-10T12:30:23.743 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../012-python3-ceph-argparse_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:23.744 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.746 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../012-python3-ceph-argparse_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:23.746 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.758 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cephfs. 2026-03-10T12:30:23.762 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-cephfs. 2026-03-10T12:30:23.763 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../013-python3-cephfs_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.764 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.767 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../013-python3-cephfs_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.769 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.778 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-10T12:30:23.781 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../014-python3-ceph-common_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:23.782 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.788 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-10T12:30:23.793 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../014-python3-ceph-common_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:23.794 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.799 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-10T12:30:23.803 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../015-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-10T12:30:23.804 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T12:30:23.815 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-10T12:30:23.820 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../015-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-10T12:30:23.822 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T12:30:23.823 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-prettytable. 2026-03-10T12:30:23.829 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../016-python3-prettytable_2.5.0-2_all.deb ... 2026-03-10T12:30:23.830 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-10T12:30:23.841 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-prettytable. 2026-03-10T12:30:23.847 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../016-python3-prettytable_2.5.0-2_all.deb ... 2026-03-10T12:30:23.848 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rbd. 2026-03-10T12:30:23.848 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-10T12:30:23.854 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../017-python3-rbd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.855 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.865 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-rbd. 2026-03-10T12:30:23.870 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../017-python3-rbd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:23.871 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:23.878 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-10T12:30:23.885 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../018-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-10T12:30:23.886 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T12:30:23.899 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-10T12:30:23.905 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../018-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-10T12:30:23.906 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T12:30:23.910 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libreadline-dev:amd64. 2026-03-10T12:30:23.917 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../019-libreadline-dev_8.1.2-1_amd64.deb ... 2026-03-10T12:30:23.918 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T12:30:23.930 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libreadline-dev:amd64. 2026-03-10T12:30:23.935 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../019-libreadline-dev_8.1.2-1_amd64.deb ... 2026-03-10T12:30:23.936 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T12:30:23.939 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package liblua5.3-dev:amd64. 2026-03-10T12:30:23.944 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../020-liblua5.3-dev_5.3.6-1build1_amd64.deb ... 2026-03-10T12:30:23.945 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T12:30:23.956 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package liblua5.3-dev:amd64. 2026-03-10T12:30:23.963 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../020-liblua5.3-dev_5.3.6-1build1_amd64.deb ... 2026-03-10T12:30:23.964 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T12:30:23.964 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package lua5.1. 2026-03-10T12:30:23.970 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../021-lua5.1_5.1.5-8.1build4_amd64.deb ... 2026-03-10T12:30:23.972 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking lua5.1 (5.1.5-8.1build4) ... 2026-03-10T12:30:23.985 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package lua5.1. 2026-03-10T12:30:23.990 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package lua-any. 2026-03-10T12:30:23.991 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../021-lua5.1_5.1.5-8.1build4_amd64.deb ... 2026-03-10T12:30:23.992 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking lua5.1 (5.1.5-8.1build4) ... 2026-03-10T12:30:23.996 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../022-lua-any_27ubuntu1_all.deb ... 2026-03-10T12:30:23.997 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking lua-any (27ubuntu1) ... 2026-03-10T12:30:24.010 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package zip. 2026-03-10T12:30:24.014 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package lua-any. 2026-03-10T12:30:24.016 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../023-zip_3.0-12build2_amd64.deb ... 2026-03-10T12:30:24.017 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking zip (3.0-12build2) ... 2026-03-10T12:30:24.020 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../022-lua-any_27ubuntu1_all.deb ... 2026-03-10T12:30:24.021 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking lua-any (27ubuntu1) ... 2026-03-10T12:30:24.037 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package zip. 2026-03-10T12:30:24.038 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package unzip. 2026-03-10T12:30:24.043 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../023-zip_3.0-12build2_amd64.deb ... 2026-03-10T12:30:24.044 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking zip (3.0-12build2) ... 2026-03-10T12:30:24.044 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../024-unzip_6.0-26ubuntu3.2_amd64.deb ... 2026-03-10T12:30:24.046 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking unzip (6.0-26ubuntu3.2) ... 2026-03-10T12:30:24.066 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package unzip. 2026-03-10T12:30:24.069 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package luarocks. 2026-03-10T12:30:24.072 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../024-unzip_6.0-26ubuntu3.2_amd64.deb ... 2026-03-10T12:30:24.073 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking unzip (6.0-26ubuntu3.2) ... 2026-03-10T12:30:24.075 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../025-luarocks_3.8.0+dfsg1-1_all.deb ... 2026-03-10T12:30:24.077 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking luarocks (3.8.0+dfsg1-1) ... 2026-03-10T12:30:24.095 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package luarocks. 2026-03-10T12:30:24.101 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../025-luarocks_3.8.0+dfsg1-1_all.deb ... 2026-03-10T12:30:24.102 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking luarocks (3.8.0+dfsg1-1) ... 2026-03-10T12:30:24.145 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package librgw2. 2026-03-10T12:30:24.151 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../026-librgw2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:24.152 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:24.168 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package librgw2. 2026-03-10T12:30:24.176 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../026-librgw2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:24.177 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:24.318 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rgw. 2026-03-10T12:30:24.321 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../027-python3-rgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:24.322 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:24.331 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-rgw. 2026-03-10T12:30:24.332 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../027-python3-rgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:24.333 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:24.346 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-10T12:30:24.349 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../028-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-10T12:30:24.351 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T12:30:24.352 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-10T12:30:24.358 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../028-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-10T12:30:24.359 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T12:30:24.369 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libradosstriper1. 2026-03-10T12:30:24.375 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../029-libradosstriper1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:24.376 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:24.377 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libradosstriper1. 2026-03-10T12:30:24.383 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../029-libradosstriper1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:24.386 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:24.402 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-common. 2026-03-10T12:30:24.408 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../030-ceph-common_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:24.410 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:24.413 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-common. 2026-03-10T12:30:24.418 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../030-ceph-common_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:24.419 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:25.691 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-base. 2026-03-10T12:30:25.692 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-base. 2026-03-10T12:30:25.697 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../031-ceph-base_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:25.697 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../031-ceph-base_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:25.701 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:25.703 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:25.877 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-10T12:30:25.878 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-10T12:30:25.883 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../032-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-10T12:30:25.883 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../032-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-10T12:30:25.883 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-10T12:30:25.884 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-10T12:30:25.897 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-cheroot. 2026-03-10T12:30:25.903 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cheroot. 2026-03-10T12:30:25.903 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../033-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-10T12:30:25.904 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T12:30:25.910 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../033-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-10T12:30:25.912 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T12:30:25.926 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-10T12:30:25.931 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../034-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-10T12:30:25.944 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-10T12:30:25.956 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-10T12:30:25.963 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../034-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-10T12:30:25.964 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-10T12:30:25.965 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-10T12:30:25.969 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../035-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-10T12:30:25.970 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-10T12:30:25.980 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-10T12:30:25.985 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-10T12:30:25.988 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../035-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-10T12:30:25.989 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../036-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-10T12:30:25.989 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-10T12:30:25.990 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-10T12:30:26.005 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-10T12:30:26.007 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-tempora. 2026-03-10T12:30:26.012 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../036-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-10T12:30:26.013 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../037-python3-tempora_4.1.2-1_all.deb ... 2026-03-10T12:30:26.013 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-10T12:30:26.014 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-10T12:30:26.044 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-portend. 2026-03-10T12:30:26.044 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-tempora. 2026-03-10T12:30:26.049 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../038-python3-portend_3.0.0-1_all.deb ... 2026-03-10T12:30:26.050 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-10T12:30:26.050 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../037-python3-tempora_4.1.2-1_all.deb ... 2026-03-10T12:30:26.051 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-10T12:30:26.067 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-10T12:30:26.068 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-portend. 2026-03-10T12:30:26.071 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../039-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-10T12:30:26.072 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-10T12:30:26.074 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../038-python3-portend_3.0.0-1_all.deb ... 2026-03-10T12:30:26.075 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-10T12:30:26.089 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-10T12:30:26.090 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-10T12:30:26.094 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../039-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-10T12:30:26.095 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-10T12:30:26.095 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../040-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-10T12:30:26.096 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-10T12:30:26.111 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-10T12:30:26.116 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../040-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-10T12:30:26.117 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-10T12:30:26.128 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-natsort. 2026-03-10T12:30:26.133 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../041-python3-natsort_8.0.2-1_all.deb ... 2026-03-10T12:30:26.135 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-10T12:30:26.146 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-natsort. 2026-03-10T12:30:26.151 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../041-python3-natsort_8.0.2-1_all.deb ... 2026-03-10T12:30:26.152 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-10T12:30:26.153 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-logutils. 2026-03-10T12:30:26.158 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../042-python3-logutils_0.3.3-8_all.deb ... 2026-03-10T12:30:26.159 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-logutils (0.3.3-8) ... 2026-03-10T12:30:26.172 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-logutils. 2026-03-10T12:30:26.173 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-mako. 2026-03-10T12:30:26.178 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../042-python3-logutils_0.3.3-8_all.deb ... 2026-03-10T12:30:26.178 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../043-python3-mako_1.1.3+ds1-2ubuntu0.1_all.deb ... 2026-03-10T12:30:26.179 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T12:30:26.179 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-logutils (0.3.3-8) ... 2026-03-10T12:30:26.199 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-mako. 2026-03-10T12:30:26.199 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-simplegeneric. 2026-03-10T12:30:26.203 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../044-python3-simplegeneric_0.8.1-3_all.deb ... 2026-03-10T12:30:26.204 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-simplegeneric (0.8.1-3) ... 2026-03-10T12:30:26.204 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../043-python3-mako_1.1.3+ds1-2ubuntu0.1_all.deb ... 2026-03-10T12:30:26.205 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T12:30:26.217 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-singledispatch. 2026-03-10T12:30:26.223 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../045-python3-singledispatch_3.4.0.3-3_all.deb ... 2026-03-10T12:30:26.225 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-singledispatch (3.4.0.3-3) ... 2026-03-10T12:30:26.226 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-simplegeneric. 2026-03-10T12:30:26.232 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../044-python3-simplegeneric_0.8.1-3_all.deb ... 2026-03-10T12:30:26.233 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-simplegeneric (0.8.1-3) ... 2026-03-10T12:30:26.241 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-webob. 2026-03-10T12:30:26.248 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../046-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-10T12:30:26.248 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-singledispatch. 2026-03-10T12:30:26.249 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T12:30:26.255 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../045-python3-singledispatch_3.4.0.3-3_all.deb ... 2026-03-10T12:30:26.256 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-singledispatch (3.4.0.3-3) ... 2026-03-10T12:30:26.271 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-waitress. 2026-03-10T12:30:26.272 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-webob. 2026-03-10T12:30:26.278 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../047-python3-waitress_1.4.4-1.1ubuntu1.1_all.deb ... 2026-03-10T12:30:26.278 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../046-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-10T12:30:26.279 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T12:30:26.280 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T12:30:26.302 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-tempita. 2026-03-10T12:30:26.303 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-waitress. 2026-03-10T12:30:26.308 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../048-python3-tempita_0.5.2-6ubuntu1_all.deb ... 2026-03-10T12:30:26.310 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../047-python3-waitress_1.4.4-1.1ubuntu1.1_all.deb ... 2026-03-10T12:30:26.310 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T12:30:26.312 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T12:30:26.326 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-paste. 2026-03-10T12:30:26.330 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-tempita. 2026-03-10T12:30:26.332 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../049-python3-paste_3.5.0+dfsg1-1_all.deb ... 2026-03-10T12:30:26.333 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T12:30:26.336 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../048-python3-tempita_0.5.2-6ubuntu1_all.deb ... 2026-03-10T12:30:26.337 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T12:30:26.356 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-paste. 2026-03-10T12:30:26.362 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../049-python3-paste_3.5.0+dfsg1-1_all.deb ... 2026-03-10T12:30:26.363 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T12:30:26.368 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python-pastedeploy-tpl. 2026-03-10T12:30:26.374 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../050-python-pastedeploy-tpl_2.1.1-1_all.deb ... 2026-03-10T12:30:26.375 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T12:30:26.399 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-pastedeploy. 2026-03-10T12:30:26.401 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python-pastedeploy-tpl. 2026-03-10T12:30:26.404 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../051-python3-pastedeploy_2.1.1-1_all.deb ... 2026-03-10T12:30:26.405 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-pastedeploy (2.1.1-1) ... 2026-03-10T12:30:26.408 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../050-python-pastedeploy-tpl_2.1.1-1_all.deb ... 2026-03-10T12:30:26.409 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T12:30:26.422 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-webtest. 2026-03-10T12:30:26.426 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pastedeploy. 2026-03-10T12:30:26.428 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../052-python3-webtest_2.0.35-1_all.deb ... 2026-03-10T12:30:26.429 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-webtest (2.0.35-1) ... 2026-03-10T12:30:26.431 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../051-python3-pastedeploy_2.1.1-1_all.deb ... 2026-03-10T12:30:26.432 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pastedeploy (2.1.1-1) ... 2026-03-10T12:30:26.445 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-pecan. 2026-03-10T12:30:26.446 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-webtest. 2026-03-10T12:30:26.451 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../052-python3-webtest_2.0.35-1_all.deb ... 2026-03-10T12:30:26.451 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../053-python3-pecan_1.3.3-4ubuntu2_all.deb ... 2026-03-10T12:30:26.451 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-webtest (2.0.35-1) ... 2026-03-10T12:30:26.452 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T12:30:26.468 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pecan. 2026-03-10T12:30:26.472 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../053-python3-pecan_1.3.3-4ubuntu2_all.deb ... 2026-03-10T12:30:26.472 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T12:30:26.485 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-werkzeug. 2026-03-10T12:30:26.491 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../054-python3-werkzeug_2.0.2+dfsg1-1ubuntu0.22.04.3_all.deb ... 2026-03-10T12:30:26.492 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T12:30:26.505 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-werkzeug. 2026-03-10T12:30:26.509 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../054-python3-werkzeug_2.0.2+dfsg1-1ubuntu0.22.04.3_all.deb ... 2026-03-10T12:30:26.548 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T12:30:26.563 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-10T12:30:26.567 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../055-ceph-mgr-modules-core_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:26.568 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.576 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-10T12:30:26.581 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../055-ceph-mgr-modules-core_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:26.582 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.607 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-10T12:30:26.612 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../056-libsqlite3-mod-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:26.613 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.619 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-10T12:30:26.624 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../056-libsqlite3-mod-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:26.625 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.628 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-mgr. 2026-03-10T12:30:26.634 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../057-ceph-mgr_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:26.635 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.640 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr. 2026-03-10T12:30:26.645 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../057-ceph-mgr_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:26.646 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.670 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-mon. 2026-03-10T12:30:26.674 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mon. 2026-03-10T12:30:26.675 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../058-ceph-mon_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:26.677 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.679 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../058-ceph-mon_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:26.680 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.805 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-10T12:30:26.805 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-10T12:30:26.811 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../059-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-10T12:30:26.812 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../059-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-10T12:30:26.812 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T12:30:26.813 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T12:30:26.833 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-osd. 2026-03-10T12:30:26.837 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-osd. 2026-03-10T12:30:26.839 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../060-ceph-osd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:26.840 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:26.842 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../060-ceph-osd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:26.843 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.180 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph. 2026-03-10T12:30:27.185 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../061-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:27.221 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.235 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-fuse. 2026-03-10T12:30:27.238 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph. 2026-03-10T12:30:27.239 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../061-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:27.240 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.240 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../062-ceph-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:27.241 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.270 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-fuse. 2026-03-10T12:30:27.274 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-mds. 2026-03-10T12:30:27.276 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../062-ceph-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:27.277 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.279 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../063-ceph-mds_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:27.280 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.326 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mds. 2026-03-10T12:30:27.328 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package cephadm. 2026-03-10T12:30:27.330 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../063-ceph-mds_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:27.331 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.333 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../064-cephadm_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:27.334 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.354 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-10T12:30:27.360 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../065-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T12:30:27.367 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T12:30:27.382 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package cephadm. 2026-03-10T12:30:27.388 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../064-cephadm_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:27.390 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.394 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-10T12:30:27.400 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../066-ceph-mgr-cephadm_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:27.401 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.409 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-10T12:30:27.415 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../065-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T12:30:27.416 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T12:30:27.431 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-10T12:30:27.437 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../067-python3-repoze.lru_0.7-2_all.deb ... 2026-03-10T12:30:27.438 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-10T12:30:27.450 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-10T12:30:27.456 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../066-ceph-mgr-cephadm_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:27.457 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.457 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-routes. 2026-03-10T12:30:27.463 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../068-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-10T12:30:27.465 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T12:30:27.482 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-10T12:30:27.487 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../067-python3-repoze.lru_0.7-2_all.deb ... 2026-03-10T12:30:27.488 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-10T12:30:27.493 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-10T12:30:27.499 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../069-ceph-mgr-dashboard_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:27.501 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:27.514 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-routes. 2026-03-10T12:30:27.519 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../068-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-10T12:30:27.520 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T12:30:27.546 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-10T12:30:27.548 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../069-ceph-mgr-dashboard_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:27.549 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:28.014 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-10T12:30:28.020 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../070-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-10T12:30:28.021 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T12:30:28.028 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-10T12:30:28.034 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../070-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-10T12:30:28.035 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T12:30:28.143 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-joblib. 2026-03-10T12:30:28.143 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-joblib. 2026-03-10T12:30:28.149 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../071-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-10T12:30:28.149 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../071-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-10T12:30:28.150 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T12:30:28.150 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T12:30:28.188 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-10T12:30:28.191 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-10T12:30:28.193 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../072-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-10T12:30:28.194 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-10T12:30:28.194 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../072-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-10T12:30:28.195 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-10T12:30:28.211 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-sklearn. 2026-03-10T12:30:28.213 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-sklearn. 2026-03-10T12:30:28.217 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../073-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-10T12:30:28.218 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T12:30:28.219 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../073-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-10T12:30:28.220 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T12:30:28.367 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-10T12:30:28.368 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-10T12:30:28.369 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../074-ceph-mgr-diskprediction-local_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:28.370 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:28.374 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../074-ceph-mgr-diskprediction-local_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:28.375 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:28.743 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cachetools. 2026-03-10T12:30:28.747 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-cachetools. 2026-03-10T12:30:28.749 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../075-python3-cachetools_5.0.0-1_all.deb ... 2026-03-10T12:30:28.750 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-10T12:30:28.753 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../075-python3-cachetools_5.0.0-1_all.deb ... 2026-03-10T12:30:28.754 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-10T12:30:28.768 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rsa. 2026-03-10T12:30:28.771 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-rsa. 2026-03-10T12:30:28.774 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../076-python3-rsa_4.8-1_all.deb ... 2026-03-10T12:30:28.775 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-10T12:30:28.778 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../076-python3-rsa_4.8-1_all.deb ... 2026-03-10T12:30:28.779 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-10T12:30:28.796 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-google-auth. 2026-03-10T12:30:28.800 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-google-auth. 2026-03-10T12:30:28.802 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../077-python3-google-auth_1.5.1-3_all.deb ... 2026-03-10T12:30:28.806 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../077-python3-google-auth_1.5.1-3_all.deb ... 2026-03-10T12:30:28.806 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-10T12:30:28.807 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-10T12:30:28.826 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-10T12:30:28.829 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-10T12:30:28.834 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../078-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-10T12:30:28.835 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T12:30:28.835 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../078-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-10T12:30:28.836 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T12:30:28.854 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-websocket. 2026-03-10T12:30:28.856 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-websocket. 2026-03-10T12:30:28.860 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../079-python3-websocket_1.2.3-1_all.deb ... 2026-03-10T12:30:28.860 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-10T12:30:28.862 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../079-python3-websocket_1.2.3-1_all.deb ... 2026-03-10T12:30:28.863 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-10T12:30:28.883 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-10T12:30:28.887 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-10T12:30:28.890 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../080-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-10T12:30:28.893 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../080-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-10T12:30:28.907 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T12:30:28.907 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T12:30:29.088 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-10T12:30:29.089 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-10T12:30:29.094 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../081-ceph-mgr-k8sevents_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:29.095 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../081-ceph-mgr-k8sevents_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:29.095 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:29.096 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:29.112 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-10T12:30:29.113 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-10T12:30:29.117 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../082-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-10T12:30:29.118 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T12:30:29.119 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../082-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-10T12:30:29.121 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T12:30:29.138 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-10T12:30:29.142 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-10T12:30:29.145 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../083-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T12:30:29.146 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T12:30:29.149 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../083-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T12:30:29.150 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T12:30:29.163 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package jq. 2026-03-10T12:30:29.170 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../084-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T12:30:29.170 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package jq. 2026-03-10T12:30:29.171 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-10T12:30:29.176 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../084-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T12:30:29.177 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-10T12:30:29.188 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package socat. 2026-03-10T12:30:29.192 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package socat. 2026-03-10T12:30:29.195 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../085-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-10T12:30:29.196 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-10T12:30:29.199 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../085-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-10T12:30:29.200 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-10T12:30:29.224 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package xmlstarlet. 2026-03-10T12:30:29.226 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package xmlstarlet. 2026-03-10T12:30:29.230 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../086-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-10T12:30:29.231 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../086-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-10T12:30:29.231 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-10T12:30:29.232 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-10T12:30:29.280 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-test. 2026-03-10T12:30:29.280 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-test. 2026-03-10T12:30:29.286 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../087-ceph-test_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:29.287 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../087-ceph-test_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:29.287 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:29.287 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:30.974 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-volume. 2026-03-10T12:30:30.979 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../088-ceph-volume_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:30.982 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:30.982 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package ceph-volume. 2026-03-10T12:30:30.988 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../088-ceph-volume_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T12:30:30.989 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:31.012 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-10T12:30:31.018 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../089-libcephfs-dev_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:31.019 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:31.020 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-10T12:30:31.026 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../089-libcephfs-dev_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:31.027 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:31.037 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package lua-socket:amd64. 2026-03-10T12:30:31.044 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../090-lua-socket_3.0~rc1+git+ac3201d-6_amd64.deb ... 2026-03-10T12:30:31.045 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package lua-socket:amd64. 2026-03-10T12:30:31.045 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T12:30:31.051 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../090-lua-socket_3.0~rc1+git+ac3201d-6_amd64.deb ... 2026-03-10T12:30:31.052 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T12:30:31.072 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package lua-sec:amd64. 2026-03-10T12:30:31.077 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../091-lua-sec_1.0.2-1_amd64.deb ... 2026-03-10T12:30:31.078 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package lua-sec:amd64. 2026-03-10T12:30:31.079 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking lua-sec:amd64 (1.0.2-1) ... 2026-03-10T12:30:31.084 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../091-lua-sec_1.0.2-1_amd64.deb ... 2026-03-10T12:30:31.085 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking lua-sec:amd64 (1.0.2-1) ... 2026-03-10T12:30:31.100 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package nvme-cli. 2026-03-10T12:30:31.105 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package nvme-cli. 2026-03-10T12:30:31.106 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../092-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-10T12:30:31.107 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T12:30:31.112 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../092-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-10T12:30:31.112 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T12:30:31.154 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package pkg-config. 2026-03-10T12:30:31.154 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package pkg-config. 2026-03-10T12:30:31.160 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../093-pkg-config_0.29.2-1ubuntu3_amd64.deb ... 2026-03-10T12:30:31.161 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T12:30:31.162 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../093-pkg-config_0.29.2-1ubuntu3_amd64.deb ... 2026-03-10T12:30:31.163 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T12:30:31.179 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-10T12:30:31.180 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-10T12:30:31.186 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../094-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T12:30:31.187 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../094-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T12:30:31.187 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T12:30:31.188 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T12:30:31.242 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-10T12:30:31.243 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-10T12:30:31.249 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../095-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-10T12:30:31.250 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-10T12:30:31.252 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../095-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-10T12:30:31.253 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-10T12:30:31.266 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-pastescript. 2026-03-10T12:30:31.270 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pastescript. 2026-03-10T12:30:31.271 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../096-python3-pastescript_2.0.2-4_all.deb ... 2026-03-10T12:30:31.272 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-pastescript (2.0.2-4) ... 2026-03-10T12:30:31.276 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../096-python3-pastescript_2.0.2-4_all.deb ... 2026-03-10T12:30:31.277 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pastescript (2.0.2-4) ... 2026-03-10T12:30:31.295 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-pluggy. 2026-03-10T12:30:31.301 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../097-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-10T12:30:31.301 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pluggy. 2026-03-10T12:30:31.302 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-10T12:30:31.308 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../097-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-10T12:30:31.309 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-10T12:30:31.324 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-psutil. 2026-03-10T12:30:31.326 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-psutil. 2026-03-10T12:30:31.330 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../098-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-10T12:30:31.332 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../098-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-10T12:30:31.333 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-10T12:30:31.333 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-10T12:30:31.359 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-py. 2026-03-10T12:30:31.359 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-py. 2026-03-10T12:30:31.365 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../099-python3-py_1.10.0-1_all.deb ... 2026-03-10T12:30:31.366 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../099-python3-py_1.10.0-1_all.deb ... 2026-03-10T12:30:31.366 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-10T12:30:31.371 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-10T12:30:31.395 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-pygments. 2026-03-10T12:30:31.399 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pygments. 2026-03-10T12:30:31.402 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../100-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-10T12:30:31.403 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T12:30:31.405 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../100-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-10T12:30:31.416 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T12:30:31.473 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-pyinotify. 2026-03-10T12:30:31.479 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../101-python3-pyinotify_0.9.6-1.3_all.deb ... 2026-03-10T12:30:31.480 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-pyinotify (0.9.6-1.3) ... 2026-03-10T12:30:31.493 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pyinotify. 2026-03-10T12:30:31.500 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../101-python3-pyinotify_0.9.6-1.3_all.deb ... 2026-03-10T12:30:31.501 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pyinotify (0.9.6-1.3) ... 2026-03-10T12:30:31.505 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-toml. 2026-03-10T12:30:31.512 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../102-python3-toml_0.10.2-1_all.deb ... 2026-03-10T12:30:31.512 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-10T12:30:31.523 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-toml. 2026-03-10T12:30:31.524 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../102-python3-toml_0.10.2-1_all.deb ... 2026-03-10T12:30:31.525 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-10T12:30:31.530 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-pytest. 2026-03-10T12:30:31.536 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../103-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-10T12:30:31.537 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T12:30:31.546 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pytest. 2026-03-10T12:30:31.552 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../103-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-10T12:30:31.553 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T12:30:31.570 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-simplejson. 2026-03-10T12:30:31.576 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../104-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-10T12:30:31.577 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-10T12:30:31.585 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-simplejson. 2026-03-10T12:30:31.591 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../104-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-10T12:30:31.592 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-10T12:30:31.600 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-10T12:30:31.605 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../105-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-10T12:30:31.606 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-10T12:30:31.614 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-10T12:30:31.620 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../105-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-10T12:30:31.621 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-10T12:30:31.736 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package radosgw. 2026-03-10T12:30:31.742 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../106-radosgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:31.743 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:31.748 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package radosgw. 2026-03-10T12:30:31.755 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../106-radosgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:31.756 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:32.047 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package rbd-fuse. 2026-03-10T12:30:32.050 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package rbd-fuse. 2026-03-10T12:30:32.051 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../107-rbd-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:32.052 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:32.056 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../107-rbd-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T12:30:32.057 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:32.079 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package smartmontools. 2026-03-10T12:30:32.079 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package smartmontools. 2026-03-10T12:30:32.079 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../108-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-10T12:30:32.081 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../108-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-10T12:30:32.087 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T12:30:32.089 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T12:30:32.137 INFO:teuthology.orchestra.run.vm05.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T12:30:32.142 INFO:teuthology.orchestra.run.vm01.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T12:30:32.379 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-10T12:30:32.379 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-10T12:30:32.407 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-10T12:30:32.407 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-10T12:30:32.763 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-10T12:30:32.815 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-10T12:30:32.837 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T12:30:32.840 INFO:teuthology.orchestra.run.vm05.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T12:30:32.885 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T12:30:32.888 INFO:teuthology.orchestra.run.vm01.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T12:30:32.911 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T12:30:32.953 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T12:30:33.140 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-10T12:30:33.199 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-10T12:30:33.543 INFO:teuthology.orchestra.run.vm05.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-10T12:30:33.550 INFO:teuthology.orchestra.run.vm05.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-10T12:30:33.552 INFO:teuthology.orchestra.run.vm05.stdout:Setting up cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:33.596 INFO:teuthology.orchestra.run.vm05.stdout:Adding system user cephadm....done 2026-03-10T12:30:33.598 INFO:teuthology.orchestra.run.vm01.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-10T12:30:33.605 INFO:teuthology.orchestra.run.vm01.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-10T12:30:33.605 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T12:30:33.606 INFO:teuthology.orchestra.run.vm01.stdout:Setting up cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:33.793 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-10T12:30:33.858 INFO:teuthology.orchestra.run.vm01.stdout:Adding system user cephadm....done 2026-03-10T12:30:33.867 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T12:30:33.913 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T12:30:33.915 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-10T12:30:33.940 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-10T12:30:33.980 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-10T12:30:34.006 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T12:30:34.010 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-10T12:30:34.049 INFO:teuthology.orchestra.run.vm05.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T12:30:34.051 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-10T12:30:34.075 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-10T12:30:34.139 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T12:30:34.142 INFO:teuthology.orchestra.run.vm01.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T12:30:34.145 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-10T12:30:34.239 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T12:30:34.264 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-10T12:30:34.332 INFO:teuthology.orchestra.run.vm05.stdout:Setting up unzip (6.0-26ubuntu3.2) ... 2026-03-10T12:30:34.340 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pyinotify (0.9.6-1.3) ... 2026-03-10T12:30:34.363 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-10T12:30:34.410 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-10T12:30:34.434 INFO:teuthology.orchestra.run.vm01.stdout:Setting up unzip (6.0-26ubuntu3.2) ... 2026-03-10T12:30:34.444 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-pyinotify (0.9.6-1.3) ... 2026-03-10T12:30:34.484 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:34.519 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-10T12:30:34.558 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T12:30:34.561 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-10T12:30:34.564 INFO:teuthology.orchestra.run.vm05.stdout:Setting up lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T12:30:34.567 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T12:30:34.569 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T12:30:34.572 INFO:teuthology.orchestra.run.vm05.stdout:Setting up lua5.1 (5.1.5-8.1build4) ... 2026-03-10T12:30:34.576 INFO:teuthology.orchestra.run.vm05.stdout:update-alternatives: using /usr/bin/lua5.1 to provide /usr/bin/lua (lua-interpreter) in auto mode 2026-03-10T12:30:34.578 INFO:teuthology.orchestra.run.vm05.stdout:update-alternatives: using /usr/bin/luac5.1 to provide /usr/bin/luac (lua-compiler) in auto mode 2026-03-10T12:30:34.580 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T12:30:34.583 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-10T12:30:34.587 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:34.657 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T12:30:34.660 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-10T12:30:34.663 INFO:teuthology.orchestra.run.vm01.stdout:Setting up lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T12:30:34.666 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T12:30:34.668 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T12:30:34.670 INFO:teuthology.orchestra.run.vm01.stdout:Setting up lua5.1 (5.1.5-8.1build4) ... 2026-03-10T12:30:34.674 INFO:teuthology.orchestra.run.vm01.stdout:update-alternatives: using /usr/bin/lua5.1 to provide /usr/bin/lua (lua-interpreter) in auto mode 2026-03-10T12:30:34.677 INFO:teuthology.orchestra.run.vm01.stdout:update-alternatives: using /usr/bin/luac5.1 to provide /usr/bin/luac (lua-compiler) in auto mode 2026-03-10T12:30:34.679 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T12:30:34.681 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-10T12:30:34.708 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-10T12:30:34.781 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T12:30:34.814 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-10T12:30:34.858 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-10T12:30:34.888 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T12:30:34.938 INFO:teuthology.orchestra.run.vm05.stdout:Setting up zip (3.0-12build2) ... 2026-03-10T12:30:34.940 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T12:30:34.961 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-10T12:30:35.041 INFO:teuthology.orchestra.run.vm01.stdout:Setting up zip (3.0-12build2) ... 2026-03-10T12:30:35.043 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T12:30:35.224 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T12:30:35.303 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T12:30:35.306 INFO:teuthology.orchestra.run.vm05.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-10T12:30:35.308 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T12:30:35.318 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T12:30:35.396 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T12:30:35.413 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T12:30:35.414 INFO:teuthology.orchestra.run.vm01.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-10T12:30:35.417 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T12:30:35.511 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T12:30:35.549 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T12:30:35.650 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T12:30:35.680 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T12:30:35.773 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T12:30:35.787 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T12:30:35.877 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T12:30:35.895 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-10T12:30:35.957 INFO:teuthology.orchestra.run.vm05.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-10T12:30:35.959 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:35.992 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-10T12:30:36.054 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T12:30:36.060 INFO:teuthology.orchestra.run.vm01.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-10T12:30:36.062 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:36.154 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T12:30:36.648 INFO:teuthology.orchestra.run.vm05.stdout:Setting up pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T12:30:36.756 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:36.756 INFO:teuthology.orchestra.run.vm01.stdout:Setting up pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T12:30:36.761 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-10T12:30:36.777 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:36.783 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-10T12:30:36.836 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T12:30:36.838 INFO:teuthology.orchestra.run.vm05.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-10T12:30:36.841 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-10T12:30:36.854 INFO:teuthology.orchestra.run.vm01.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T12:30:36.857 INFO:teuthology.orchestra.run.vm01.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-10T12:30:36.859 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-10T12:30:36.911 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-10T12:30:36.929 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-10T12:30:36.976 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:36.978 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-10T12:30:37.000 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:37.003 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-10T12:30:37.048 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-singledispatch (3.4.0.3-3) ... 2026-03-10T12:30:37.073 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-singledispatch (3.4.0.3-3) ... 2026-03-10T12:30:37.113 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-logutils (0.3.3-8) ... 2026-03-10T12:30:37.141 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-logutils (0.3.3-8) ... 2026-03-10T12:30:37.184 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-10T12:30:37.212 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-10T12:30:37.256 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-simplegeneric (0.8.1-3) ... 2026-03-10T12:30:37.283 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-simplegeneric (0.8.1-3) ... 2026-03-10T12:30:37.324 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-10T12:30:37.349 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-10T12:30:37.398 INFO:teuthology.orchestra.run.vm05.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T12:30:37.400 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-10T12:30:37.419 INFO:teuthology.orchestra.run.vm01.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T12:30:37.421 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-10T12:30:37.478 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T12:30:37.480 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T12:30:37.500 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T12:30:37.503 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T12:30:37.553 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T12:30:37.573 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T12:30:37.639 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T12:30:37.661 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T12:30:37.733 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-10T12:30:37.756 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-10T12:30:37.801 INFO:teuthology.orchestra.run.vm05.stdout:Setting up liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T12:30:37.804 INFO:teuthology.orchestra.run.vm05.stdout:Setting up lua-sec:amd64 (1.0.2-1) ... 2026-03-10T12:30:37.806 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T12:30:37.808 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T12:30:37.826 INFO:teuthology.orchestra.run.vm01.stdout:Setting up liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T12:30:37.829 INFO:teuthology.orchestra.run.vm01.stdout:Setting up lua-sec:amd64 (1.0.2-1) ... 2026-03-10T12:30:37.831 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T12:30:37.834 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T12:30:37.944 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pastedeploy (2.1.1-1) ... 2026-03-10T12:30:37.973 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-pastedeploy (2.1.1-1) ... 2026-03-10T12:30:38.015 INFO:teuthology.orchestra.run.vm05.stdout:Setting up lua-any (27ubuntu1) ... 2026-03-10T12:30:38.018 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-10T12:30:38.046 INFO:teuthology.orchestra.run.vm01.stdout:Setting up lua-any (27ubuntu1) ... 2026-03-10T12:30:38.048 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-10T12:30:38.088 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:38.090 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-10T12:30:38.119 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:30:38.124 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-10T12:30:38.167 INFO:teuthology.orchestra.run.vm05.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-10T12:30:38.169 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-webtest (2.0.35-1) ... 2026-03-10T12:30:38.204 INFO:teuthology.orchestra.run.vm01.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-10T12:30:38.207 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-webtest (2.0.35-1) ... 2026-03-10T12:30:38.267 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-10T12:30:38.280 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-10T12:30:38.406 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pastescript (2.0.2-4) ... 2026-03-10T12:30:38.412 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-pastescript (2.0.2-4) ... 2026-03-10T12:30:38.494 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T12:30:38.496 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T12:30:38.610 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T12:30:38.612 INFO:teuthology.orchestra.run.vm01.stdout:Setting up librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:38.615 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:38.618 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T12:30:38.621 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T12:30:38.626 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:38.629 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:38.631 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T12:30:39.214 INFO:teuthology.orchestra.run.vm01.stdout:Setting up luarocks (3.8.0+dfsg1-1) ... 2026-03-10T12:30:39.221 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.227 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.227 INFO:teuthology.orchestra.run.vm05.stdout:Setting up luarocks (3.8.0+dfsg1-1) ... 2026-03-10T12:30:39.229 INFO:teuthology.orchestra.run.vm01.stdout:Setting up librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.232 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.234 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.234 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.237 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.240 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.242 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.244 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.295 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T12:30:39.295 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T12:30:39.304 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T12:30:39.305 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T12:30:39.656 INFO:teuthology.orchestra.run.vm01.stdout:Setting up libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.659 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.664 INFO:teuthology.orchestra.run.vm01.stdout:Setting up librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.667 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.669 INFO:teuthology.orchestra.run.vm01.stdout:Setting up rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.672 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.675 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.677 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.683 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.686 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.688 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.690 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.693 INFO:teuthology.orchestra.run.vm05.stdout:Setting up rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.695 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.698 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.700 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:39.710 INFO:teuthology.orchestra.run.vm01.stdout:Adding group ceph....done 2026-03-10T12:30:39.734 INFO:teuthology.orchestra.run.vm05.stdout:Adding group ceph....done 2026-03-10T12:30:39.744 INFO:teuthology.orchestra.run.vm01.stdout:Adding system user ceph....done 2026-03-10T12:30:39.753 INFO:teuthology.orchestra.run.vm01.stdout:Setting system user ceph properties....done 2026-03-10T12:30:39.757 INFO:teuthology.orchestra.run.vm01.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-10T12:30:39.773 INFO:teuthology.orchestra.run.vm05.stdout:Adding system user ceph....done 2026-03-10T12:30:39.782 INFO:teuthology.orchestra.run.vm05.stdout:Setting system user ceph properties....done 2026-03-10T12:30:39.786 INFO:teuthology.orchestra.run.vm05.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-10T12:30:39.823 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-10T12:30:39.852 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-10T12:30:40.077 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-10T12:30:40.080 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-10T12:30:40.421 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:40.423 INFO:teuthology.orchestra.run.vm05.stdout:Setting up radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:40.466 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:40.468 INFO:teuthology.orchestra.run.vm01.stdout:Setting up radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:40.685 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:30:40.685 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:30:40.689 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:30:40.690 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:30:41.052 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:41.063 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:41.147 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-10T12:30:41.147 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-10T12:30:41.497 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:41.549 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:41.562 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T12:30:41.563 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T12:30:41.617 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T12:30:41.617 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T12:30:41.924 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:41.984 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T12:30:41.984 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T12:30:42.023 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:42.084 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T12:30:42.084 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T12:30:42.365 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:42.442 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T12:30:42.442 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T12:30:42.484 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:42.559 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T12:30:42.560 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T12:30:42.812 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:42.814 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:42.828 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:42.889 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T12:30:42.889 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T12:30:42.933 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:42.935 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:42.949 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.011 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T12:30:43.011 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T12:30:43.260 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.273 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.275 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.288 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.379 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.399 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.450 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T12:30:43.451 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.459 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:30:43.466 INFO:teuthology.orchestra.run.vm01.stdout:Setting up ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:30:43.474 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:30:43.555 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for install-info (6.8-4build1) ... 2026-03-10T12:30:43.585 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T12:30:43.593 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:30:43.615 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:30:43.724 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for install-info (6.8-4build1) ... 2026-03-10T12:30:43.899 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:43.900 INFO:teuthology.orchestra.run.vm05.stdout:Running kernel seems to be up-to-date. 2026-03-10T12:30:43.900 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:43.900 INFO:teuthology.orchestra.run.vm05.stdout:Services to be restarted: 2026-03-10T12:30:43.906 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart packagekit.service 2026-03-10T12:30:43.909 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout:Service restarts being deferred: 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart unattended-upgrades.service 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout:No containers need to be restarted. 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout:No user sessions are running outdated binaries. 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:43.910 INFO:teuthology.orchestra.run.vm05.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T12:30:44.061 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:44.062 INFO:teuthology.orchestra.run.vm01.stdout:Running kernel seems to be up-to-date. 2026-03-10T12:30:44.062 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:44.062 INFO:teuthology.orchestra.run.vm01.stdout:Services to be restarted: 2026-03-10T12:30:44.067 INFO:teuthology.orchestra.run.vm01.stdout: systemctl restart packagekit.service 2026-03-10T12:30:44.069 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout:Service restarts being deferred: 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout: systemctl restart unattended-upgrades.service 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout:No containers need to be restarted. 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout:No user sessions are running outdated binaries. 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:44.070 INFO:teuthology.orchestra.run.vm01.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T12:30:44.863 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:30:44.866 DEBUG:teuthology.orchestra.run.vm05:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-xmltodict python3-jmespath 2026-03-10T12:30:44.947 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:30:45.020 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:30:45.023 DEBUG:teuthology.orchestra.run.vm01:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-xmltodict python3-jmespath 2026-03-10T12:30:45.082 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:30:45.082 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:30:45.098 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:30:45.247 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:30:45.247 INFO:teuthology.orchestra.run.vm05.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:30:45.248 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T12:30:45.248 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:30:45.263 INFO:teuthology.orchestra.run.vm05.stdout:The following NEW packages will be installed: 2026-03-10T12:30:45.263 INFO:teuthology.orchestra.run.vm05.stdout: python3-jmespath python3-xmltodict 2026-03-10T12:30:45.298 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:30:45.298 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:30:45.351 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 2 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:30:45.351 INFO:teuthology.orchestra.run.vm05.stdout:Need to get 34.3 kB of archives. 2026-03-10T12:30:45.351 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 146 kB of additional disk space will be used. 2026-03-10T12:30:45.351 INFO:teuthology.orchestra.run.vm05.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-10T12:30:45.368 INFO:teuthology.orchestra.run.vm05.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-10T12:30:45.441 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:30:45.441 INFO:teuthology.orchestra.run.vm01.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:30:45.441 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T12:30:45.441 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:30:45.454 INFO:teuthology.orchestra.run.vm01.stdout:The following NEW packages will be installed: 2026-03-10T12:30:45.454 INFO:teuthology.orchestra.run.vm01.stdout: python3-jmespath python3-xmltodict 2026-03-10T12:30:45.561 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 34.3 kB in 0s (339 kB/s) 2026-03-10T12:30:45.575 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jmespath. 2026-03-10T12:30:45.603 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118577 files and directories currently installed.) 2026-03-10T12:30:45.605 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-10T12:30:45.606 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-10T12:30:45.623 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-10T12:30:45.629 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-10T12:30:45.629 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-10T12:30:45.658 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-10T12:30:45.722 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-10T12:30:45.900 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 2 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:30:45.900 INFO:teuthology.orchestra.run.vm01.stdout:Need to get 34.3 kB of archives. 2026-03-10T12:30:45.900 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 146 kB of additional disk space will be used. 2026-03-10T12:30:45.900 INFO:teuthology.orchestra.run.vm01.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-10T12:30:46.041 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:46.041 INFO:teuthology.orchestra.run.vm05.stdout:Running kernel seems to be up-to-date. 2026-03-10T12:30:46.041 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:46.041 INFO:teuthology.orchestra.run.vm05.stdout:Services to be restarted: 2026-03-10T12:30:46.046 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart packagekit.service 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout:Service restarts being deferred: 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart unattended-upgrades.service 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout:No containers need to be restarted. 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout:No user sessions are running outdated binaries. 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:30:46.049 INFO:teuthology.orchestra.run.vm05.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T12:30:46.112 INFO:teuthology.orchestra.run.vm01.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-10T12:30:46.309 INFO:teuthology.orchestra.run.vm01.stdout:Fetched 34.3 kB in 1s (51.7 kB/s) 2026-03-10T12:30:46.324 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-jmespath. 2026-03-10T12:30:46.354 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118577 files and directories currently installed.) 2026-03-10T12:30:46.357 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-10T12:30:46.358 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-10T12:30:46.374 INFO:teuthology.orchestra.run.vm01.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-10T12:30:46.379 INFO:teuthology.orchestra.run.vm01.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-10T12:30:46.380 INFO:teuthology.orchestra.run.vm01.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-10T12:30:46.410 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-10T12:30:46.484 INFO:teuthology.orchestra.run.vm01.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-10T12:30:46.831 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:46.831 INFO:teuthology.orchestra.run.vm01.stdout:Running kernel seems to be up-to-date. 2026-03-10T12:30:46.831 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:46.831 INFO:teuthology.orchestra.run.vm01.stdout:Services to be restarted: 2026-03-10T12:30:46.836 INFO:teuthology.orchestra.run.vm01.stdout: systemctl restart packagekit.service 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout:Service restarts being deferred: 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout: systemctl restart unattended-upgrades.service 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout:No containers need to be restarted. 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout:No user sessions are running outdated binaries. 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:30:46.839 INFO:teuthology.orchestra.run.vm01.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T12:30:46.976 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:30:46.979 DEBUG:teuthology.parallel:result is None 2026-03-10T12:30:47.764 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:30:47.767 DEBUG:teuthology.parallel:result is None 2026-03-10T12:30:47.767 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:30:48.370 DEBUG:teuthology.orchestra.run.vm01:> dpkg-query -W -f '${Version}' ceph 2026-03-10T12:30:48.379 INFO:teuthology.orchestra.run.vm01.stdout:19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:30:48.379 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:30:48.379 INFO:teuthology.task.install:The correct ceph version 19.2.3-678-ge911bdeb-1jammy is installed. 2026-03-10T12:30:48.380 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:30:48.956 DEBUG:teuthology.orchestra.run.vm05:> dpkg-query -W -f '${Version}' ceph 2026-03-10T12:30:48.965 INFO:teuthology.orchestra.run.vm05.stdout:19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:30:48.965 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T12:30:48.965 INFO:teuthology.task.install:The correct ceph version 19.2.3-678-ge911bdeb-1jammy is installed. 2026-03-10T12:30:48.966 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-10T12:30:48.966 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:30:48.966 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T12:30:48.973 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:30:48.973 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T12:30:49.014 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-10T12:30:49.014 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:30:49.014 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T12:30:49.022 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T12:30:49.070 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:30:49.070 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T12:30:49.079 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T12:30:49.130 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-10T12:30:49.130 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:30:49.130 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T12:30:49.138 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T12:30:49.187 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:30:49.187 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T12:30:49.194 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T12:30:49.246 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-10T12:30:49.246 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:30:49.246 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T12:30:49.254 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T12:30:49.302 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:30:49.302 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T12:30:49.310 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T12:30:49.363 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-10T12:30:49.408 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 3}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': True}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T12:30:49.408 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:30:49.408 INFO:tasks.cephadm:Cluster fsid is f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:30:49.408 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-10T12:30:49.408 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.101', 'mon.b': '192.168.123.105'} 2026-03-10T12:30:49.408 INFO:tasks.cephadm:First mon is mon.a on vm01 2026-03-10T12:30:49.408 INFO:tasks.cephadm:First mgr is a 2026-03-10T12:30:49.408 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-10T12:30:49.408 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-10T12:30:49.416 DEBUG:teuthology.orchestra.run.vm05:> sudo hostname $(hostname -s) 2026-03-10T12:30:49.425 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-10T12:30:49.425 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:30:50.094 INFO:tasks.cephadm:builder_project result: [{'url': 'https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/', 'chacra_url': 'https://1.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'ubuntu', 'distro_version': '22.04', 'distro_codename': 'jammy', 'modified': '2026-02-25 19:37:07.680480', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678-ge911bdeb-1jammy', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.98+toko08', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-10T12:30:50.683 INFO:tasks.util.chacra:got chacra host 1.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=ubuntu%2F22.04%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:30:50.684 INFO:tasks.cephadm:Discovered cachra url: https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm 2026-03-10T12:30:50.684 INFO:tasks.cephadm:Downloading cephadm from url: https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm 2026-03-10T12:30:50.684 DEBUG:teuthology.orchestra.run.vm01:> curl --silent -L https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T12:30:52.072 INFO:teuthology.orchestra.run.vm01.stdout:-rw-rw-r-- 1 ubuntu ubuntu 795696 Mar 10 12:30 /home/ubuntu/cephtest/cephadm 2026-03-10T12:30:52.072 DEBUG:teuthology.orchestra.run.vm05:> curl --silent -L https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T12:30:53.387 INFO:teuthology.orchestra.run.vm05.stdout:-rw-rw-r-- 1 ubuntu ubuntu 795696 Mar 10 12:30 /home/ubuntu/cephtest/cephadm 2026-03-10T12:30:53.387 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T12:30:53.391 DEBUG:teuthology.orchestra.run.vm05:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T12:30:53.398 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-10T12:30:53.398 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T12:30:53.436 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T12:30:53.526 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T12:30:53.536 INFO:teuthology.orchestra.run.vm05.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T12:31:54.815 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-10T12:31:54.815 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T12:31:54.815 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T12:31:54.815 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-10T12:31:54.815 INFO:teuthology.orchestra.run.vm01.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T12:31:54.815 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-10T12:31:54.815 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-10T12:32:33.120 INFO:teuthology.orchestra.run.vm05.stdout:{ 2026-03-10T12:32:33.120 INFO:teuthology.orchestra.run.vm05.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T12:32:33.120 INFO:teuthology.orchestra.run.vm05.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T12:32:33.120 INFO:teuthology.orchestra.run.vm05.stdout: "repo_digests": [ 2026-03-10T12:32:33.120 INFO:teuthology.orchestra.run.vm05.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T12:32:33.120 INFO:teuthology.orchestra.run.vm05.stdout: ] 2026-03-10T12:32:33.120 INFO:teuthology.orchestra.run.vm05.stdout:} 2026-03-10T12:32:33.140 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-10T12:32:33.147 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /etc/ceph 2026-03-10T12:32:33.155 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-10T12:32:33.197 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 777 /etc/ceph 2026-03-10T12:32:33.206 INFO:tasks.cephadm:Writing seed config... 2026-03-10T12:32:33.206 INFO:tasks.cephadm: override: [global] mon election default strategy = 3 2026-03-10T12:32:33.206 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-10T12:32:33.206 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-10T12:32:33.207 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = True 2026-03-10T12:32:33.207 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-10T12:32:33.207 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-10T12:32:33.207 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-10T12:32:33.207 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-10T12:32:33.207 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-10T12:32:33.207 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-10T12:32:33.207 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:32:33.207 DEBUG:teuthology.orchestra.run.vm01:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-10T12:32:33.241 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = f864ef90-1c7c-11f1-946a-0b343fff599a mon election default strategy = 3 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = True [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-10T12:32:33.241 DEBUG:teuthology.orchestra.run.vm01:mon.a> sudo journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a.service 2026-03-10T12:32:33.284 DEBUG:teuthology.orchestra.run.vm01:mgr.a> sudo journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.a.service 2026-03-10T12:32:33.328 INFO:tasks.cephadm:Bootstrapping... 2026-03-10T12:32:33.328 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid f864ef90-1c7c-11f1-946a-0b343fff599a --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.101 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:32:33.465 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-10T12:32:33.465 INFO:teuthology.orchestra.run.vm01.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', 'f864ef90-1c7c-11f1-946a-0b343fff599a', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.101', '--skip-admin-label'] 2026-03-10T12:32:33.465 INFO:teuthology.orchestra.run.vm01.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-10T12:32:33.466 INFO:teuthology.orchestra.run.vm01.stdout:Verifying podman|docker is present... 2026-03-10T12:32:33.466 INFO:teuthology.orchestra.run.vm01.stdout:Verifying lvm2 is present... 2026-03-10T12:32:33.466 INFO:teuthology.orchestra.run.vm01.stdout:Verifying time synchronization is in place... 2026-03-10T12:32:33.469 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T12:32:33.469 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T12:32:33.472 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T12:32:33.472 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T12:32:33.475 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-10T12:32:33.475 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-10T12:32:33.478 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-10T12:32:33.478 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T12:32:33.481 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-10T12:32:33.481 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout masked 2026-03-10T12:32:33.484 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-10T12:32:33.484 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T12:32:33.486 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-10T12:32:33.487 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-10T12:32:33.490 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-10T12:32:33.490 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T12:32:33.494 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-10T12:32:33.496 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-10T12:32:33.496 INFO:teuthology.orchestra.run.vm01.stdout:Unit ntp.service is enabled and running 2026-03-10T12:32:33.496 INFO:teuthology.orchestra.run.vm01.stdout:Repeating the final host check... 2026-03-10T12:32:33.496 INFO:teuthology.orchestra.run.vm01.stdout:docker (/usr/bin/docker) is present 2026-03-10T12:32:33.496 INFO:teuthology.orchestra.run.vm01.stdout:systemctl is present 2026-03-10T12:32:33.496 INFO:teuthology.orchestra.run.vm01.stdout:lvcreate is present 2026-03-10T12:32:33.499 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T12:32:33.499 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T12:32:33.501 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T12:32:33.501 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T12:32:33.504 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-10T12:32:33.504 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-10T12:32:33.507 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-10T12:32:33.507 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T12:32:33.509 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-10T12:32:33.509 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout masked 2026-03-10T12:32:33.511 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-10T12:32:33.512 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T12:32:33.514 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-10T12:32:33.514 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-10T12:32:33.517 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-10T12:32:33.517 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T12:32:33.520 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-10T12:32:33.523 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-10T12:32:33.523 INFO:teuthology.orchestra.run.vm01.stdout:Unit ntp.service is enabled and running 2026-03-10T12:32:33.523 INFO:teuthology.orchestra.run.vm01.stdout:Host looks OK 2026-03-10T12:32:33.523 INFO:teuthology.orchestra.run.vm01.stdout:Cluster fsid: f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:33.523 INFO:teuthology.orchestra.run.vm01.stdout:Acquiring lock 140013819578928 on /run/cephadm/f864ef90-1c7c-11f1-946a-0b343fff599a.lock 2026-03-10T12:32:33.523 INFO:teuthology.orchestra.run.vm01.stdout:Lock 140013819578928 acquired on /run/cephadm/f864ef90-1c7c-11f1-946a-0b343fff599a.lock 2026-03-10T12:32:33.523 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 3300 ... 2026-03-10T12:32:33.524 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 6789 ... 2026-03-10T12:32:33.524 INFO:teuthology.orchestra.run.vm01.stdout:Base mon IP(s) is [192.168.123.101:3300, 192.168.123.101:6789], mon addrv is [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T12:32:33.525 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.101 metric 100 2026-03-10T12:32:33.525 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-10T12:32:33.525 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout 192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.101 metric 100 2026-03-10T12:32:33.525 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout 192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.101 metric 100 2026-03-10T12:32:33.527 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-10T12:32:33.527 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout fe80::/64 dev ens3 proto kernel metric 256 pref medium 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout 2: ens3: mtu 1500 state UP qlen 1000 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout inet6 fe80::5055:ff:fe00:1/64 scope link 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.1/32` 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.1/32` 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24', '192.168.123.1/32', '192.168.123.1/32'] 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-10T12:32:33.529 INFO:teuthology.orchestra.run.vm01.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T12:32:34.657 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/docker: stdout e911bdebe5c8faa3800735d1568fcdca65db60df: Pulling from ceph-ci/ceph 2026-03-10T12:32:34.657 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/docker: stdout Digest: sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T12:32:34.657 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/docker: stdout Status: Image is up to date for quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:32:34.657 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/docker: stdout quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:32:34.824 INFO:teuthology.orchestra.run.vm01.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T12:32:34.824 INFO:teuthology.orchestra.run.vm01.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T12:32:34.824 INFO:teuthology.orchestra.run.vm01.stdout:Extracting ceph user uid/gid from container image... 2026-03-10T12:32:34.932 INFO:teuthology.orchestra.run.vm01.stdout:stat: stdout 167 167 2026-03-10T12:32:34.932 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial keys... 2026-03-10T12:32:35.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQBjD7BpfQcQARAAhk8q8cGEcXOTmR5iRmhV3g== 2026-03-10T12:32:35.166 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQBjD7BpRNQfCBAA/m21XfD5dYVbJdxvrBhFzw== 2026-03-10T12:32:35.282 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQBjD7BpSHMnDxAAjLpTOlKNaH2cCsq6Ry1H0w== 2026-03-10T12:32:35.283 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial monmap... 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:monmaptool for a [v2:192.168.123.101:3300,v1:192.168.123.101:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:setting min_mon_release = quincy 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: set fsid to f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:32:35.398 INFO:teuthology.orchestra.run.vm01.stdout:Creating mon... 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 1 imported monmap: 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr epoch 0 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr last_changed 2026-03-10T12:32:35.364821+0000 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr min_mon_release 17 (quincy) 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr election_strategy: 1 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 0 /usr/bin/ceph-mon: set fsid to f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:35.546 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Git sha 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: DB SUMMARY 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: DB Session ID: 00XGS5APJQEM7Q9U3JL9 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 0, files: 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.create_if_missing: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.env: 0x55f007261dc0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.info_log: 0x55f01622cda0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.statistics: (nil) 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.use_fsync: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.db_log_dir: 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.wal_dir: 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.write_buffer_manager: 0x55f0162235e0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.unordered_write: 0 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T12:32:35.547 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.row_cache: None 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.wal_filter: None 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.wal_compression: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_open_files: -1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Compression algorithms supported: 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: kZSTD supported: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.492+0000 7f586b7b7d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: [db/db_impl/db_impl_open.cc:317] Creating manifest 1 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.merge_operator: 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_filter: None 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55f01621f520) 2026-03-10T12:32:35.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr pin_top_level_index_and_filter: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr index_type: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr data_block_index_type: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr index_shortening: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr data_block_hash_table_util_ratio: 0.750000 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr checksum: 4 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr no_block_cache: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr block_cache: 0x55f016245350 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr block_cache_name: BinnedLRUCache 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr block_cache_options: 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr capacity : 536870912 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr num_shard_bits : 4 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr strict_capacity_limit : 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr high_pri_pool_ratio: 0.000 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr block_cache_compressed: (nil) 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr persistent_cache: (nil) 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr block_size: 4096 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr block_size_deviation: 10 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr block_restart_interval: 16 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr index_block_restart_interval: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr metadata_block_size: 4096 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr partition_filters: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr use_delta_encoding: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr filter_policy: bloomfilter 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr whole_key_filtering: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr verify_compression: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr read_amp_bytes_per_bit: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr format_version: 5 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr enable_index_compression: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr block_align: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr max_auto_readahead_size: 262144 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr prepopulate_block_cache: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr initial_auto_readahead_size: 8192 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr num_file_reads_for_auto_readahead: 2 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression: NoCompression 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.num_levels: 7 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T12:32:35.549 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.ttl: 2592000 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T12:32:35.550 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 succeeded,manifest_file_number is 1, next_file_number is 3, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.496+0000 7f586b7b7d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 38f355c1-e07f-4226-a178-26f487d5fa48 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.500+0000 7f586b7b7d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 5 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.500+0000 7f586b7b7d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55f016246e00 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.500+0000 7f586b7b7d80 4 rocksdb: DB pointer 0x55f01632a000 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.500+0000 7f5862f41640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.500+0000 7f5862f41640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr ** DB Stats ** 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr AddFile(Total Files): cumulative 0, interval 0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr AddFile(Keys): cumulative 0, interval 0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Block cache BinnedLRUCache@0x55f016245350#7 capacity: 512.00 MB usage: 0.00 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 9e-06 secs_since: 0 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr Block cache entry stats(count,size,portion): Misc(1,0.00 KB,0%) 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr ** File Read Latency Histogram By Level [default] ** 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.504+0000 7f586b7b7d80 4 rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.504+0000 7f586b7b7d80 4 rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T12:32:35.504+0000 7f586b7b7d80 0 /usr/bin/ceph-mon: created monfs at /var/lib/ceph/mon/ceph-a for mon.a 2026-03-10T12:32:35.551 INFO:teuthology.orchestra.run.vm01.stdout:create mon.a on 2026-03-10T12:32:35.734 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Removed /etc/systemd/system/multi-user.target.wants/ceph.target. 2026-03-10T12:32:35.903 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-10T12:32:36.177 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a.target → /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a.target. 2026-03-10T12:32:36.177 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a.target → /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a.target. 2026-03-10T12:32:36.376 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:32:36.385 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a 2026-03-10T12:32:36.385 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a.service: Unit ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a.service not loaded. 2026-03-10T12:32:36.586 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a.target.wants/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a.service → /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service. 2026-03-10T12:32:36.595 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T12:32:36.595 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T12:32:36.595 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon to start... 2026-03-10T12:32:36.595 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon... 2026-03-10T12:32:36.653 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:32:36.653 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:32:36.653 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:32:36.653 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:32:36.653 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 systemd[1]: Started Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout cluster: 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout id: f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout services: 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.064457s) 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout data: 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pgs: 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:mon is available 2026-03-10T12:32:36.873 INFO:teuthology.orchestra.run.vm01.stdout:Assimilating anything we can from ceph.conf... 2026-03-10T12:32:36.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 bash[20237]: cluster 2026-03-10T12:32:36.746413+0000 mon.a (mon.0) 0 : cluster [INF] mkfs f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:36.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 bash[20237]: cluster 2026-03-10T12:32:36.746413+0000 mon.a (mon.0) 0 : cluster [INF] mkfs f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:36.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 bash[20237]: cluster 2026-03-10T12:32:36.741242+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:32:36.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:36 vm01 bash[20237]: cluster 2026-03-10T12:32:36.741242+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = True 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T12:32:37.153 INFO:teuthology.orchestra.run.vm01.stdout:Generating new minimal ceph.conf... 2026-03-10T12:32:37.388 INFO:teuthology.orchestra.run.vm01.stdout:Restarting the monitor... 2026-03-10T12:32:37.667 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 systemd[1]: Stopping Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:32:37.667 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20237]: debug 2026-03-10T12:32:37.428+0000 7ff9b6a25640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-10T12:32:37.667 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20237]: debug 2026-03-10T12:32:37.428+0000 7ff9b6a25640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-10T12:32:37.733 INFO:teuthology.orchestra.run.vm01.stdout:Setting public_network to 192.168.123.1/32,192.168.123.0/24 in mon config section 2026-03-10T12:32:37.917 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20631]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-a 2026-03-10T12:32:37.917 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 systemd[1]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a.service: Deactivated successfully. 2026-03-10T12:32:37.917 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 systemd[1]: Stopped Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:32:37.917 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 systemd[1]: Started Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:32:38.074 INFO:teuthology.orchestra.run.vm01.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-10T12:32:38.074 INFO:teuthology.orchestra.run.vm01.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:32:38.074 INFO:teuthology.orchestra.run.vm01.stdout:Creating mgr... 2026-03-10T12:32:38.075 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-10T12:32:38.075 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 0 pidfile_write: ignore empty --pid-file 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 0 load: jerasure load: lrc 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Git sha 0 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: DB SUMMARY 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: DB Session ID: QTFE08QO8GOGJSQ95CBK 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T12:32:38.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75507 ; 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.env: 0x56337d381dc0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.info_log: 0x563395c6e700 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.statistics: (nil) 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.use_fsync: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.db_log_dir: 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.wal_dir: 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.write_buffer_manager: 0x563395c73900 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T12:32:38.212 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.unordered_write: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.row_cache: None 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.wal_filter: None 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.wal_compression: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_open_files: -1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Compression algorithms supported: 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: kZSTD supported: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.merge_operator: 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_filter: None 2026-03-10T12:32:38.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x563395c6e640) 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: cache_index_and_filter_blocks: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: pin_top_level_index_and_filter: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: index_type: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: data_block_index_type: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: index_shortening: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: checksum: 4 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: no_block_cache: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: block_cache: 0x563395c95350 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: block_cache_name: BinnedLRUCache 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: block_cache_options: 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: capacity : 536870912 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: num_shard_bits : 4 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: strict_capacity_limit : 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: high_pri_pool_ratio: 0.000 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: block_cache_compressed: (nil) 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: persistent_cache: (nil) 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: block_size: 4096 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: block_size_deviation: 10 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: block_restart_interval: 16 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: index_block_restart_interval: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: metadata_block_size: 4096 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: partition_filters: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: use_delta_encoding: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: filter_policy: bloomfilter 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: whole_key_filtering: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: verify_compression: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: read_amp_bytes_per_bit: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: format_version: 5 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: enable_index_compression: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: block_align: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: max_auto_readahead_size: 262144 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: prepopulate_block_cache: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: initial_auto_readahead_size: 8192 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: num_file_reads_for_auto_readahead: 2 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression: NoCompression 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.num_levels: 7 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T12:32:38.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.ttl: 2592000 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T12:32:38.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.948+0000 7f32c84aed80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.968+0000 7f32c84aed80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.968+0000 7f32c84aed80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.968+0000 7f32c84aed80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 38f355c1-e07f-4226-a178-26f487d5fa48 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.968+0000 7f32c84aed80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773145957973594, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.968+0000 7f32c84aed80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.972+0000 7f32c84aed80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773145957975731, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72588, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 70867, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9705, "raw_average_key_size": 49, "raw_value_size": 65346, "raw_average_value_size": 333, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773145957, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "38f355c1-e07f-4226-a178-26f487d5fa48", "db_session_id": "QTFE08QO8GOGJSQ95CBK", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.972+0000 7f32c84aed80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773145957975800, "job": 1, "event": "recovery_finished"} 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.972+0000 7f32c84aed80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.972+0000 7f32c84aed80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.972+0000 7f32c84aed80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x563395c96e00 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.972+0000 7f32c84aed80 4 rocksdb: DB pointer 0x563395dac000 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.976+0000 7f32be278640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: debug 2026-03-10T12:32:37.976+0000 7f32be278640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: ** DB Stats ** 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: ** Compaction Stats [default] ** 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: L0 2/0 72.74 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 35.7 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Sum 2/0 72.74 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 35.7 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 35.7 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: ** Compaction Stats [default] ** 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 35.7 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: AddFile(Keys): cumulative 0, interval 0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Cumulative compaction: 0.00 GB write, 2.82 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Interval compaction: 0.00 GB write, 2.82 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Block cache BinnedLRUCache@0x563395c95350#7 capacity: 512.00 MB usage: 1.06 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.4e-05 secs_since: 0 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: Block cache entry stats(count,size,portion): FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(1,0.00 KB,0%) 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:37 vm01 bash[20716]: ** File Read Latency Histogram By Level [default] ** 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985791+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985791+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985834+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T12:32:38.216 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985834+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985840+0000 mon.a (mon.0) 3 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985840+0000 mon.a (mon.0) 3 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985843+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T12:32:35.364821+0000 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985843+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T12:32:35.364821+0000 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985851+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985851+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985855+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985855+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985860+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985860+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985863+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.985863+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.986107+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.986107+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.986121+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.986121+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.986638+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: cluster 2026-03-10T12:32:37.986638+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: audit 2026-03-10T12:32:38.032213+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.101:0/586937715' entity='client.admin' 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20716]: audit 2026-03-10T12:32:38.032213+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.101:0/586937715' entity='client.admin' 2026-03-10T12:32:38.217 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:32:38.268 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.a 2026-03-10T12:32:38.268 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.a.service: Unit ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.a.service not loaded. 2026-03-10T12:32:38.458 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a.target.wants/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.a.service → /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service. 2026-03-10T12:32:38.462 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:38 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:32:38.462 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:38 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:32:38.467 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T12:32:38.467 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T12:32:38.467 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T12:32:38.467 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-10T12:32:38.467 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr to start... 2026-03-10T12:32:38.467 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr... 2026-03-10T12:32:38.707 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:38.707 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:32:38.707 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "f864ef90-1c7c-11f1-946a-0b343fff599a", 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T12:32:38.708 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T12:32:36:745933+0000", 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T12:32:36.746668+0000", 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:32:38.709 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (1/15)... 2026-03-10T12:32:38.728 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:38 vm01 systemd[1]: Started Ceph mgr.a for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:32:38.728 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20987]: debug 2026-03-10T12:32:38.668+0000 7fc4d6188140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T12:32:39.041 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20987]: debug 2026-03-10T12:32:38.724+0000 7fc4d6188140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T12:32:39.041 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:38 vm01 bash[20987]: debug 2026-03-10T12:32:38.852+0000 7fc4d6188140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T12:32:39.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20716]: audit 2026-03-10T12:32:38.654260+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.101:0/4165219441' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:32:39.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20716]: audit 2026-03-10T12:32:38.654260+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.101:0/4165219441' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:32:39.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20987]: debug 2026-03-10T12:32:39.160+0000 7fc4d6188140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T12:32:39.949 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20987]: debug 2026-03-10T12:32:39.652+0000 7fc4d6188140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T12:32:39.949 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20987]: debug 2026-03-10T12:32:39.740+0000 7fc4d6188140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T12:32:39.949 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20987]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T12:32:39.949 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20987]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T12:32:39.949 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20987]: from numpy import show_config as show_numpy_config 2026-03-10T12:32:39.949 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:39 vm01 bash[20987]: debug 2026-03-10T12:32:39.864+0000 7fc4d6188140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T12:32:40.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.004+0000 7fc4d6188140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T12:32:40.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.044+0000 7fc4d6188140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T12:32:40.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.084+0000 7fc4d6188140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T12:32:40.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.128+0000 7fc4d6188140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T12:32:40.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.180+0000 7fc4d6188140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T12:32:40.886 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.612+0000 7fc4d6188140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T12:32:40.886 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.652+0000 7fc4d6188140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T12:32:40.886 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.696+0000 7fc4d6188140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "f864ef90-1c7c-11f1-946a-0b343fff599a", 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 2, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T12:32:40.952 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T12:32:36:745933+0000", 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T12:32:36.746668+0000", 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:32:40.953 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (2/15)... 2026-03-10T12:32:41.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20716]: audit 2026-03-10T12:32:40.886421+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.101:0/670267606' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:32:41.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20716]: audit 2026-03-10T12:32:40.886421+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.101:0/670267606' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:32:41.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.880+0000 7fc4d6188140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T12:32:41.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.924+0000 7fc4d6188140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T12:32:41.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:40 vm01 bash[20987]: debug 2026-03-10T12:32:40.972+0000 7fc4d6188140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T12:32:41.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20987]: debug 2026-03-10T12:32:41.084+0000 7fc4d6188140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:32:41.647 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20987]: debug 2026-03-10T12:32:41.244+0000 7fc4d6188140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T12:32:41.648 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20987]: debug 2026-03-10T12:32:41.412+0000 7fc4d6188140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T12:32:41.648 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20987]: debug 2026-03-10T12:32:41.448+0000 7fc4d6188140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T12:32:41.648 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20987]: debug 2026-03-10T12:32:41.492+0000 7fc4d6188140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T12:32:41.898 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20987]: debug 2026-03-10T12:32:41.640+0000 7fc4d6188140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:32:42.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20987]: debug 2026-03-10T12:32:41.892+0000 7fc4d6188140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T12:32:42.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: cluster 2026-03-10T12:32:41.899560+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T12:32:42.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: cluster 2026-03-10T12:32:41.899560+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T12:32:42.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: cluster 2026-03-10T12:32:41.904872+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00542696s) 2026-03-10T12:32:42.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: cluster 2026-03-10T12:32:41.904872+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00542696s) 2026-03-10T12:32:42.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.907537+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.907537+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.908004+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.908004+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.908414+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.908414+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.908904+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.908904+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.910052+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.910052+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: cluster 2026-03-10T12:32:41.916614+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: cluster 2026-03-10T12:32:41.916614+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.929855+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.929855+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.931607+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.931607+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.935085+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.935085+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.936540+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:32:42.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:41 vm01 bash[20716]: audit 2026-03-10T12:32:41.936540+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:32:43.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:42 vm01 bash[20716]: audit 2026-03-10T12:32:41.944600+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:32:43.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:42 vm01 bash[20716]: audit 2026-03-10T12:32:41.944600+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:32:43.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:42 vm01 bash[20716]: cluster 2026-03-10T12:32:42.909841+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0104s) 2026-03-10T12:32:43.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:42 vm01 bash[20716]: cluster 2026-03-10T12:32:42.909841+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0104s) 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "f864ef90-1c7c-11f1-946a-0b343fff599a", 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T12:32:43.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T12:32:36:745933+0000", 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T12:32:36.746668+0000", 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:32:43.267 INFO:teuthology.orchestra.run.vm01.stdout:mgr is available 2026-03-10T12:32:43.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:43.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T12:32:43.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:43.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T12:32:43.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T12:32:43.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T12:32:43.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T12:32:43.544 INFO:teuthology.orchestra.run.vm01.stdout:Enabling cephadm module... 2026-03-10T12:32:44.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:43 vm01 bash[20716]: audit 2026-03-10T12:32:43.224964+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.101:0/3145018730' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:32:44.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:43 vm01 bash[20716]: audit 2026-03-10T12:32:43.224964+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.101:0/3145018730' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:32:44.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:43 vm01 bash[20716]: audit 2026-03-10T12:32:43.496151+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.101:0/4287141444' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T12:32:44.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:43 vm01 bash[20716]: audit 2026-03-10T12:32:43.496151+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.101:0/4287141444' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T12:32:44.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:43 vm01 bash[20716]: audit 2026-03-10T12:32:43.498855+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.101:0/4287141444' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T12:32:44.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:43 vm01 bash[20716]: audit 2026-03-10T12:32:43.498855+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.101:0/4287141444' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T12:32:44.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:43 vm01 bash[20716]: audit 2026-03-10T12:32:43.799213+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.101:0/2672616149' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T12:32:44.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:43 vm01 bash[20716]: audit 2026-03-10T12:32:43.799213+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.101:0/2672616149' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T12:32:44.829 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:44 vm01 bash[20987]: ignoring --setuser ceph since I am not root 2026-03-10T12:32:44.829 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:44 vm01 bash[20987]: ignoring --setgroup ceph since I am not root 2026-03-10T12:32:44.829 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:44 vm01 bash[20987]: debug 2026-03-10T12:32:44.640+0000 7fb37d973140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T12:32:44.829 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:44 vm01 bash[20987]: debug 2026-03-10T12:32:44.684+0000 7fb37d973140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T12:32:45.165 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:44 vm01 bash[20987]: debug 2026-03-10T12:32:44.824+0000 7fb37d973140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T12:32:45.257 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:32:45.257 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-10T12:32:45.257 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T12:32:45.257 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-10T12:32:45.257 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T12:32:45.257 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:32:45.257 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-10T12:32:45.257 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 4... 2026-03-10T12:32:45.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20987]: debug 2026-03-10T12:32:45.160+0000 7fb37d973140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T12:32:45.752 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20987]: debug 2026-03-10T12:32:45.660+0000 7fb37d973140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T12:32:45.752 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20716]: audit 2026-03-10T12:32:44.499484+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.101:0/2672616149' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T12:32:45.752 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20716]: audit 2026-03-10T12:32:44.499484+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.101:0/2672616149' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T12:32:45.752 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20716]: cluster 2026-03-10T12:32:44.502305+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T12:32:45.752 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20716]: cluster 2026-03-10T12:32:44.502305+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T12:32:45.752 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20716]: audit 2026-03-10T12:32:44.876320+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.101:0/1608840306' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:32:45.752 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20716]: audit 2026-03-10T12:32:44.876320+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.101:0/1608840306' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:32:46.030 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20987]: debug 2026-03-10T12:32:45.748+0000 7fb37d973140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T12:32:46.030 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20987]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T12:32:46.030 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20987]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T12:32:46.030 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20987]: from numpy import show_config as show_numpy_config 2026-03-10T12:32:46.030 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:45 vm01 bash[20987]: debug 2026-03-10T12:32:45.880+0000 7fb37d973140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T12:32:46.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.024+0000 7fb37d973140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T12:32:46.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.064+0000 7fb37d973140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T12:32:46.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.100+0000 7fb37d973140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T12:32:46.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.144+0000 7fb37d973140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T12:32:46.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.196+0000 7fb37d973140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T12:32:46.904 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.628+0000 7fb37d973140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T12:32:46.904 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.672+0000 7fb37d973140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T12:32:46.904 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.716+0000 7fb37d973140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T12:32:46.904 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.856+0000 7fb37d973140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T12:32:47.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.900+0000 7fb37d973140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T12:32:47.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:46 vm01 bash[20987]: debug 2026-03-10T12:32:46.940+0000 7fb37d973140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T12:32:47.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20987]: debug 2026-03-10T12:32:47.048+0000 7fb37d973140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:32:47.476 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20987]: debug 2026-03-10T12:32:47.208+0000 7fb37d973140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T12:32:47.476 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20987]: debug 2026-03-10T12:32:47.392+0000 7fb37d973140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T12:32:47.476 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20987]: debug 2026-03-10T12:32:47.428+0000 7fb37d973140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T12:32:47.864 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20987]: debug 2026-03-10T12:32:47.472+0000 7fb37d973140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T12:32:47.864 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20987]: debug 2026-03-10T12:32:47.624+0000 7fb37d973140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:32:48.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20987]: debug 2026-03-10T12:32:47.860+0000 7fb37d973140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T12:32:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.865370+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T12:32:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.865370+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T12:32:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.865781+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T12:32:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.865781+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T12:32:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.869569+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T12:32:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.869569+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T12:32:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.869660+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00397717s) 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.869660+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00397717s) 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.872484+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.872484+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.872844+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.872844+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.874161+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.874161+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.874537+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.874537+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.874886+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.874886+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.881414+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: cluster 2026-03-10T12:32:47.881414+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.890921+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.890921+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.894237+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.894237+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.896388+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.896388+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.908480+0000 mon.a (mon.0) 49 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.908480+0000 mon.a (mon.0) 49 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.910074+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.910074+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.910409+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.910409+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.916967+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:32:48.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:47 vm01 bash[20716]: audit 2026-03-10T12:32:47.916967+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:32:48.927 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:32:48.927 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-10T12:32:48.927 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T12:32:48.927 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:32:48.927 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 4 is available 2026-03-10T12:32:48.927 INFO:teuthology.orchestra.run.vm01.stdout:Setting orchestrator backend to cephadm... 2026-03-10T12:32:49.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: cephadm 2026-03-10T12:32:47.888805+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T12:32:49.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: cephadm 2026-03-10T12:32:47.888805+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T12:32:49.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: audit 2026-03-10T12:32:48.384448+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:49.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: audit 2026-03-10T12:32:48.384448+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:49.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: audit 2026-03-10T12:32:48.387537+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:49.480 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: audit 2026-03-10T12:32:48.387537+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:49.480 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: cluster 2026-03-10T12:32:48.875057+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00936s) 2026-03-10T12:32:49.480 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: cluster 2026-03-10T12:32:48.875057+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00936s) 2026-03-10T12:32:49.480 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: audit 2026-03-10T12:32:49.208682+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:49.480 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: audit 2026-03-10T12:32:49.208682+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:49.480 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: audit 2026-03-10T12:32:49.213534+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:49.480 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20716]: audit 2026-03-10T12:32:49.213534+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:49.504 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-10T12:32:49.504 INFO:teuthology.orchestra.run.vm01.stdout:Generating ssh key... 2026-03-10T12:32:50.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: Generating public/private ed25519 key pair. 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: Your identification has been saved in /tmp/tmpjctlm_fa/key 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: Your public key has been saved in /tmp/tmpjctlm_fa/key.pub 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: The key fingerprint is: 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: SHA256:SIpTQFE621aGI+pXSFqJbYnLeHIavaw7gCLykrtts2Q ceph-f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: The key's randomart image is: 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: +--[ED25519 256]--+ 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: | .+o. | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: | +oo. | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: | o+B+ + | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: |oo*B.* . | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: |*+B.+.. S | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: |BB +. | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: |*+E. | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: |+== | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: |=Boo | 2026-03-10T12:32:50.200 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:49 vm01 bash[20987]: +----[SHA256]-----+ 2026-03-10T12:32:50.309 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINGudTQQRVjHbMwbCgXS2XGFvnPQwfk6DEZbbghsoKue ceph-f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:32:50.309 INFO:teuthology.orchestra.run.vm01.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-10T12:32:50.309 INFO:teuthology.orchestra.run.vm01.stdout:Adding key to root@localhost authorized_keys... 2026-03-10T12:32:50.309 INFO:teuthology.orchestra.run.vm01.stdout:Adding host vm01... 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:48.875909+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:48.875909+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:48.880956+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:48.880956+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.205182+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.205182+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.242908+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Bus STARTING 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.242908+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Bus STARTING 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.359671+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.359671+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.360163+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Client ('192.168.123.101', 43636) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.360163+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Client ('192.168.123.101', 43636) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.460855+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.460855+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.460894+0000 mgr.a (mgr.14118) 9 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Bus STARTED 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.460894+0000 mgr.a (mgr.14118) 9 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Bus STARTED 2026-03-10T12:32:50.546 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.461711+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:50.547 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.461711+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:50.547 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.468411+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:50.547 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.468411+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:50.547 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.871095+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:50.547 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.871095+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:50.547 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.873857+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:50.547 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:50 vm01 bash[20716]: audit 2026-03-10T12:32:49.873857+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: audit 2026-03-10T12:32:49.849178+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: audit 2026-03-10T12:32:49.849178+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.849411+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: cephadm 2026-03-10T12:32:49.849411+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: audit 2026-03-10T12:32:50.271865+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: audit 2026-03-10T12:32:50.271865+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: cluster 2026-03-10T12:32:50.389118+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: cluster 2026-03-10T12:32:50.389118+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T12:32:51.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: audit 2026-03-10T12:32:50.538622+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:51.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:51 vm01 bash[20716]: audit 2026-03-10T12:32:50.538622+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:52.409 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Added host 'vm01' with addr '192.168.123.101' 2026-03-10T12:32:52.409 INFO:teuthology.orchestra.run.vm01.stdout:Deploying unmanaged mon service... 2026-03-10T12:32:52.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:52 vm01 bash[20716]: cephadm 2026-03-10T12:32:51.088722+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm01 2026-03-10T12:32:52.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:52 vm01 bash[20716]: cephadm 2026-03-10T12:32:51.088722+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm01 2026-03-10T12:32:52.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:52 vm01 bash[20716]: audit 2026-03-10T12:32:52.335859+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:52.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:52 vm01 bash[20716]: audit 2026-03-10T12:32:52.335859+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:52.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:52 vm01 bash[20716]: audit 2026-03-10T12:32:52.338872+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:52.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:52 vm01 bash[20716]: audit 2026-03-10T12:32:52.338872+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:52.705 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-10T12:32:52.705 INFO:teuthology.orchestra.run.vm01.stdout:Deploying unmanaged mgr service... 2026-03-10T12:32:52.982 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-10T12:32:53.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:53 vm01 bash[20716]: cephadm 2026-03-10T12:32:52.336178+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm01 2026-03-10T12:32:53.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:53 vm01 bash[20716]: cephadm 2026-03-10T12:32:52.336178+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm01 2026-03-10T12:32:53.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:53 vm01 bash[20716]: audit 2026-03-10T12:32:52.667638+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:53.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:53 vm01 bash[20716]: audit 2026-03-10T12:32:52.667638+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:53.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:53 vm01 bash[20716]: audit 2026-03-10T12:32:52.941035+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:53.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:53 vm01 bash[20716]: audit 2026-03-10T12:32:52.941035+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:53.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:53 vm01 bash[20716]: audit 2026-03-10T12:32:53.205811+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.101:0/2874561222' entity='client.admin' 2026-03-10T12:32:53.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:53 vm01 bash[20716]: audit 2026-03-10T12:32:53.205811+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.101:0/2874561222' entity='client.admin' 2026-03-10T12:32:53.527 INFO:teuthology.orchestra.run.vm01.stdout:Enabling the dashboard module... 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:52.663571+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:52.663571+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: cephadm 2026-03-10T12:32:52.664464+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: cephadm 2026-03-10T12:32:52.664464+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:52.937111+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:52.937111+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: cephadm 2026-03-10T12:32:52.937809+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: cephadm 2026-03-10T12:32:52.937809+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:53.481442+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.101:0/3519582229' entity='client.admin' 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:53.481442+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.101:0/3519582229' entity='client.admin' 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:53.812170+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:53.812170+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:54.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:53.831814+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.101:0/2164443391' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T12:32:54.828 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:53.831814+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.101:0/2164443391' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T12:32:54.828 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:54.113979+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:54.828 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20716]: audit 2026-03-10T12:32:54.113979+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:32:55.151 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20987]: ignoring --setuser ceph since I am not root 2026-03-10T12:32:55.151 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20987]: ignoring --setgroup ceph since I am not root 2026-03-10T12:32:55.151 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:54 vm01 bash[20987]: debug 2026-03-10T12:32:54.956+0000 7f546d824140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T12:32:55.151 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20987]: debug 2026-03-10T12:32:55.020+0000 7f546d824140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T12:32:55.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:32:55.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-10T12:32:55.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T12:32:55.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-10T12:32:55.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T12:32:55.266 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:32:55.266 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-10T12:32:55.266 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 8... 2026-03-10T12:32:55.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20987]: debug 2026-03-10T12:32:55.144+0000 7f546d824140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T12:32:55.814 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20987]: debug 2026-03-10T12:32:55.460+0000 7f546d824140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T12:32:56.108 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20716]: audit 2026-03-10T12:32:54.812436+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.101:0/2164443391' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T12:32:56.108 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20716]: audit 2026-03-10T12:32:54.812436+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.101:0/2164443391' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T12:32:56.108 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20716]: cluster 2026-03-10T12:32:54.818334+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T12:32:56.108 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20716]: cluster 2026-03-10T12:32:54.818334+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T12:32:56.108 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20716]: audit 2026-03-10T12:32:55.211171+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.101:0/2437500333' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:32:56.108 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20716]: audit 2026-03-10T12:32:55.211171+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.101:0/2437500333' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:32:56.108 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20987]: debug 2026-03-10T12:32:55.900+0000 7f546d824140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T12:32:56.108 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:55 vm01 bash[20987]: debug 2026-03-10T12:32:55.984+0000 7f546d824140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T12:32:56.365 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T12:32:56.365 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T12:32:56.365 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: from numpy import show_config as show_numpy_config 2026-03-10T12:32:56.365 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.108+0000 7f546d824140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T12:32:56.365 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.244+0000 7f546d824140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T12:32:56.365 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.280+0000 7f546d824140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T12:32:56.365 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.316+0000 7f546d824140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T12:32:56.699 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.360+0000 7f546d824140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T12:32:56.699 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.416+0000 7f546d824140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T12:32:57.122 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.864+0000 7f546d824140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T12:32:57.122 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.900+0000 7f546d824140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T12:32:57.122 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:56 vm01 bash[20987]: debug 2026-03-10T12:32:56.936+0000 7f546d824140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T12:32:57.122 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.076+0000 7f546d824140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T12:32:57.414 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.116+0000 7f546d824140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T12:32:57.414 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.152+0000 7f546d824140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T12:32:57.414 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.260+0000 7f546d824140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:32:57.699 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.408+0000 7f546d824140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T12:32:57.699 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.572+0000 7f546d824140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T12:32:57.699 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.608+0000 7f546d824140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T12:32:57.699 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.648+0000 7f546d824140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T12:32:58.071 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:57 vm01 bash[20987]: debug 2026-03-10T12:32:57.788+0000 7f546d824140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:32:58.071 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20987]: debug 2026-03-10T12:32:58.016+0000 7f546d824140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T12:32:58.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.023943+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T12:32:58.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.023943+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T12:32:58.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.024338+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T12:32:58.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.024338+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.028461+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.028461+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.028600+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00435932s) 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.028600+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00435932s) 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.032632+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.032632+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.032978+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.032978+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.033497+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.033497+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.033879+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.033879+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.034230+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.034230+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.039219+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: cluster 2026-03-10T12:32:58.039219+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.050802+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.050802+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.063683+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:58.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:58 vm01 bash[20716]: audit 2026-03-10T12:32:58.063683+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:32:59.086 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:32:59.086 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-10T12:32:59.086 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T12:32:59.086 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:32:59.086 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 8 is available 2026-03-10T12:32:59.086 INFO:teuthology.orchestra.run.vm01.stdout:Generating a dashboard self-signed certificate... 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: audit 2026-03-10T12:32:58.076339+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: audit 2026-03-10T12:32:58.076339+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: audit 2026-03-10T12:32:58.088275+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: audit 2026-03-10T12:32:58.088275+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.711027+0000 mgr.a (mgr.14150) 1 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Bus STARTING 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.711027+0000 mgr.a (mgr.14150) 1 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Bus STARTING 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.812622+0000 mgr.a (mgr.14150) 2 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.812622+0000 mgr.a (mgr.14150) 2 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: cluster 2026-03-10T12:32:59.032279+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00804s) 2026-03-10T12:32:59.382 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:32:59 vm01 bash[20716]: cluster 2026-03-10T12:32:59.032279+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00804s) 2026-03-10T12:32:59.409 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-10T12:32:59.409 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial admin user... 2026-03-10T12:32:59.822 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$7KxL.lm3zJfTKJp2aUp4AuB9O2hVyJ9/uzyADrCBIVp50/KAKx5KC", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773145979, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-10T12:32:59.822 INFO:teuthology.orchestra.run.vm01.stdout:Fetching dashboard port number... 2026-03-10T12:33:00.139 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 8443 2026-03-10T12:33:00.139 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T12:33:00.139 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-10T12:33:00.139 INFO:teuthology.orchestra.run.vm01.stdout:Ceph Dashboard is now available at: 2026-03-10T12:33:00.139 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.140 INFO:teuthology.orchestra.run.vm01.stdout: URL: https://vm01.local:8443/ 2026-03-10T12:33:00.140 INFO:teuthology.orchestra.run.vm01.stdout: User: admin 2026-03-10T12:33:00.140 INFO:teuthology.orchestra.run.vm01.stdout: Password: vkwsd6k6v6 2026-03-10T12:33:00.140 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.140 INFO:teuthology.orchestra.run.vm01.stdout:Saving cluster configuration to /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config directory 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout:Or, if you are only running a single cluster on this host: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: ceph telemetry on 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout:For more information see: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:00.499 INFO:teuthology.orchestra.run.vm01.stdout:Bootstrap complete. 2026-03-10T12:33:00.516 INFO:tasks.cephadm:Fetching config... 2026-03-10T12:33:00.516 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:33:00.516 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-10T12:33:00.518 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-10T12:33:00.518 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:33:00.518 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-10T12:33:00.562 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-10T12:33:00.562 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:33:00.562 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/keyring of=/dev/stdout 2026-03-10T12:33:00.612 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-10T12:33:00.612 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:33:00.612 DEBUG:teuthology.orchestra.run.vm01:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-10T12:33:00.658 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-10T12:33:00.658 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINGudTQQRVjHbMwbCgXS2XGFvnPQwfk6DEZbbghsoKue ceph-f864ef90-1c7c-11f1-946a-0b343fff599a' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.921725+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.921725+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.921769+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Bus STARTED 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.921769+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Bus STARTED 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.922173+0000 mgr.a (mgr.14150) 5 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Client ('192.168.123.101', 54550) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: cephadm 2026-03-10T12:32:58.922173+0000 mgr.a (mgr.14150) 5 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Client ('192.168.123.101', 54550) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.033545+0000 mgr.a (mgr.14150) 6 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.033545+0000 mgr.a (mgr.14150) 6 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.037630+0000 mgr.a (mgr.14150) 7 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.037630+0000 mgr.a (mgr.14150) 7 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.320182+0000 mgr.a (mgr.14150) 8 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.320182+0000 mgr.a (mgr.14150) 8 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.364047+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.364047+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.367533+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.367533+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:00.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.628654+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:00.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.628654+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:00.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.783799+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:00.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:32:59.783799+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:00.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:33:00.099772+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.101:0/4077019482' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T12:33:00.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:00 vm01 bash[20716]: audit 2026-03-10T12:33:00.099772+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.101:0/4077019482' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T12:33:00.710 INFO:teuthology.orchestra.run.vm01.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINGudTQQRVjHbMwbCgXS2XGFvnPQwfk6DEZbbghsoKue ceph-f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:00.715 DEBUG:teuthology.orchestra.run.vm05:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINGudTQQRVjHbMwbCgXS2XGFvnPQwfk6DEZbbghsoKue ceph-f864ef90-1c7c-11f1-946a-0b343fff599a' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T12:33:00.726 INFO:teuthology.orchestra.run.vm05.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINGudTQQRVjHbMwbCgXS2XGFvnPQwfk6DEZbbghsoKue ceph-f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:00.731 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-10T12:33:01.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:01 vm01 bash[20716]: audit 2026-03-10T12:33:00.433950+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.101:0/1610423079' entity='client.admin' 2026-03-10T12:33:01.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:01 vm01 bash[20716]: audit 2026-03-10T12:33:00.433950+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.101:0/1610423079' entity='client.admin' 2026-03-10T12:33:01.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:01 vm01 bash[20716]: cluster 2026-03-10T12:33:00.786608+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T12:33:01.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:01 vm01 bash[20716]: cluster 2026-03-10T12:33:00.786608+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:33:04 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.052130+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.052130+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.052676+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.052676+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.055480+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.055480+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.056423+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]: dispatch 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.056423+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]: dispatch 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.057352+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]': finished 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: audit 2026-03-10T12:33:03.057352+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]': finished 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: cephadm 2026-03-10T12:33:03.604092+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm01 on vm01 2026-03-10T12:33:04.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 bash[20716]: cephadm 2026-03-10T12:33:03.604092+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm01 on vm01 2026-03-10T12:33:04.533 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:04 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:04.533 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:33:04 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:04.895 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: cluster 2026-03-10T12:33:04.106580+0000 mon.a (mon.0) 100 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: cluster 2026-03-10T12:33:04.106580+0000 mon.a (mon.0) 100 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.305899+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.305899+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.309992+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.309992+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.312984+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.312984+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.320691+0000 mon.a (mon.0) 104 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.320691+0000 mon.a (mon.0) 104 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.324338+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.324338+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.326587+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.326587+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.328762+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.328762+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.330890+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.330890+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.333743+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.333743+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.340509+0000 mon.a (mon.0) 110 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.340509+0000 mon.a (mon.0) 110 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.344156+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.344156+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.616732+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.184 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:05 vm01 bash[20716]: audit 2026-03-10T12:33:04.616732+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:05.235 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-10T12:33:05.235 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-10T12:33:06.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:06 vm01 bash[20716]: audit 2026-03-10T12:33:05.159316+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.101:0/2715616957' entity='client.admin' 2026-03-10T12:33:06.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:06 vm01 bash[20716]: audit 2026-03-10T12:33:05.159316+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.101:0/2715616957' entity='client.admin' 2026-03-10T12:33:06.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:06 vm01 bash[20716]: audit 2026-03-10T12:33:05.165650+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:06.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:06 vm01 bash[20716]: audit 2026-03-10T12:33:05.165650+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:06.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:06 vm01 bash[20716]: audit 2026-03-10T12:33:05.169704+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:06.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:06 vm01 bash[20716]: audit 2026-03-10T12:33:05.169704+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:08.905 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:33:09.183 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:09 vm01 bash[20716]: audit 2026-03-10T12:33:08.060690+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:09.183 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:09 vm01 bash[20716]: audit 2026-03-10T12:33:08.060690+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:09.271 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm05 2026-03-10T12:33:09.271 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:33:09.271 DEBUG:teuthology.orchestra.run.vm05:> dd of=/etc/ceph/ceph.conf 2026-03-10T12:33:09.274 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:33:09.274 DEBUG:teuthology.orchestra.run.vm05:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:09.317 INFO:tasks.cephadm:Adding host vm05 to orchestrator... 2026-03-10T12:33:09.317 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch host add vm05 2026-03-10T12:33:10.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.178781+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.178781+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.181662+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.181662+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.201990+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.201990+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.204886+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.204886+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.205707+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.205707+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.206748+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.206748+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.207089+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.207089+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: cephadm 2026-03-10T12:33:09.207561+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm01:/etc/ceph/ceph.conf 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: cephadm 2026-03-10T12:33:09.207561+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm01:/etc/ceph/ceph.conf 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: cephadm 2026-03-10T12:33:09.250601+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: cephadm 2026-03-10T12:33:09.250601+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: cephadm 2026-03-10T12:33:09.294165+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: cephadm 2026-03-10T12:33:09.294165+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: cephadm 2026-03-10T12:33:09.326858+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.client.admin.keyring 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: cephadm 2026-03-10T12:33:09.326858+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.client.admin.keyring 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.364569+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.364569+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.367489+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:10.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:10 vm01 bash[20716]: audit 2026-03-10T12:33:09.367489+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:14.730 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:33:15.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:15 vm01 bash[20716]: audit 2026-03-10T12:33:14.291682+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:15.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:15 vm01 bash[20716]: audit 2026-03-10T12:33:14.291682+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:15.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:15 vm01 bash[20716]: audit 2026-03-10T12:33:14.294448+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:15.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:15 vm01 bash[20716]: audit 2026-03-10T12:33:14.294448+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:16.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:16 vm01 bash[20716]: audit 2026-03-10T12:33:15.000834+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:16.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:16 vm01 bash[20716]: audit 2026-03-10T12:33:15.000834+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:16.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:16 vm01 bash[20716]: cephadm 2026-03-10T12:33:15.540621+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm05 2026-03-10T12:33:16.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:16 vm01 bash[20716]: cephadm 2026-03-10T12:33:15.540621+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm05 2026-03-10T12:33:16.800 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm05' with addr '192.168.123.105' 2026-03-10T12:33:16.872 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch host ls --format=json 2026-03-10T12:33:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:17 vm01 bash[20716]: audit 2026-03-10T12:33:16.799193+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:17 vm01 bash[20716]: audit 2026-03-10T12:33:16.799193+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:17 vm01 bash[20716]: cephadm 2026-03-10T12:33:16.799687+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm05 2026-03-10T12:33:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:17 vm01 bash[20716]: cephadm 2026-03-10T12:33:16.799687+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm05 2026-03-10T12:33:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:17 vm01 bash[20716]: audit 2026-03-10T12:33:16.799958+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:17 vm01 bash[20716]: audit 2026-03-10T12:33:16.799958+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:17 vm01 bash[20716]: audit 2026-03-10T12:33:17.089955+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:17 vm01 bash[20716]: audit 2026-03-10T12:33:17.089955+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cluster 2026-03-10T12:33:18.035857+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:19.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cluster 2026-03-10T12:33:18.035857+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:19.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.376240+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.376240+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.377999+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.377999+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.378406+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.378406+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.378981+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.378981+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.379383+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.379383+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.379940+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.379940+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.407710+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.407710+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.433978+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.433978+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.460239+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.client.admin.keyring 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.460239+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.client.admin.keyring 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.490650+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.490650+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.493028+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.493028+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.493945+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm05", "caps": []}]: dispatch 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.493945+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm05", "caps": []}]: dispatch 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.495000+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm05", "caps": []}]': finished 2026-03-10T12:33:19.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:19 vm01 bash[20716]: audit 2026-03-10T12:33:18.495000+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm05", "caps": []}]': finished 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.948412+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm05 on vm05 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: cephadm 2026-03-10T12:33:18.948412+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm05 on vm05 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.610963+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.610963+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.613484+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.613484+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.615816+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.615816+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.625000+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.625000+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.625982+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.625982+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.626315+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.626315+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.629103+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.629103+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.637560+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.637560+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.638297+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.638297+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.638681+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.638681+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.640876+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.640876+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.896139+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:20.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:20 vm01 bash[20716]: audit 2026-03-10T12:33:19.896139+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:21.487 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:33:21.726 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:21.726 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.105", "hostname": "vm05", "labels": [], "status": ""}] 2026-03-10T12:33:21.783 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-10T12:33:21.783 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph osd crush tunables default 2026-03-10T12:33:21.783 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:21 vm01 bash[20716]: cluster 2026-03-10T12:33:20.036015+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:21.783 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:21 vm01 bash[20716]: cluster 2026-03-10T12:33:20.036015+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:21.783 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:21 vm01 bash[20716]: audit 2026-03-10T12:33:20.957904+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:21.783 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:21 vm01 bash[20716]: audit 2026-03-10T12:33:20.957904+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:21.783 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:21 vm01 bash[20716]: audit 2026-03-10T12:33:20.960135+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:21.783 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:21 vm01 bash[20716]: audit 2026-03-10T12:33:20.960135+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:22.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:22 vm01 bash[20716]: audit 2026-03-10T12:33:21.725771+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:33:22.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:22 vm01 bash[20716]: audit 2026-03-10T12:33:21.725771+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:33:23.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:23 vm01 bash[20716]: cluster 2026-03-10T12:33:22.036295+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:23.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:23 vm01 bash[20716]: cluster 2026-03-10T12:33:22.036295+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:23.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:23 vm01 bash[20716]: audit 2026-03-10T12:33:23.063503+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:23.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:23 vm01 bash[20716]: audit 2026-03-10T12:33:23.063503+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:25.496 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:33:25.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:25 vm01 bash[20716]: cluster 2026-03-10T12:33:24.036499+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:25.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:25 vm01 bash[20716]: cluster 2026-03-10T12:33:24.036499+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:26.631 INFO:teuthology.orchestra.run.vm01.stderr:adjusted tunables profile to default 2026-03-10T12:33:26.693 INFO:tasks.cephadm:Adding mon.a on vm01 2026-03-10T12:33:26.693 INFO:tasks.cephadm:Adding mon.b on vm05 2026-03-10T12:33:26.693 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch apply mon '2;vm01:192.168.123.101=a;vm05:192.168.123.105=b' 2026-03-10T12:33:26.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:26 vm01 bash[20716]: audit 2026-03-10T12:33:25.773202+0000 mon.a (mon.0) 154 : audit [INF] from='client.? 192.168.123.101:0/378617481' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T12:33:26.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:26 vm01 bash[20716]: audit 2026-03-10T12:33:25.773202+0000 mon.a (mon.0) 154 : audit [INF] from='client.? 192.168.123.101:0/378617481' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T12:33:27.814 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:27.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: cluster 2026-03-10T12:33:26.036682+0000 mgr.a (mgr.14150) 29 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:27.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: cluster 2026-03-10T12:33:26.036682+0000 mgr.a (mgr.14150) 29 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:27.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: audit 2026-03-10T12:33:26.627971+0000 mon.a (mon.0) 155 : audit [INF] from='client.? 192.168.123.101:0/378617481' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T12:33:27.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: audit 2026-03-10T12:33:26.627971+0000 mon.a (mon.0) 155 : audit [INF] from='client.? 192.168.123.101:0/378617481' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T12:33:27.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: cluster 2026-03-10T12:33:26.632410+0000 mon.a (mon.0) 156 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:33:27.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: cluster 2026-03-10T12:33:26.632410+0000 mon.a (mon.0) 156 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:33:27.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: audit 2026-03-10T12:33:27.598342+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:27.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: audit 2026-03-10T12:33:27.598342+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:27.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: audit 2026-03-10T12:33:27.602006+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:27.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: audit 2026-03-10T12:33:27.602006+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:27.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: audit 2026-03-10T12:33:27.605250+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:27.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:27 vm01 bash[20716]: audit 2026-03-10T12:33:27.605250+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:28.082 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled mon update... 2026-03-10T12:33:28.155 DEBUG:teuthology.orchestra.run.vm05:mon.b> sudo journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.b.service 2026-03-10T12:33:28.155 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T12:33:28.156 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph mon dump -f json 2026-03-10T12:33:29.319 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.b/config 2026-03-10T12:33:29.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: cluster 2026-03-10T12:33:28.036934+0000 mgr.a (mgr.14150) 30 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: cluster 2026-03-10T12:33:28.036934+0000 mgr.a (mgr.14150) 30 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.077185+0000 mgr.a (mgr.14150) 31 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm01:192.168.123.101=a;vm05:192.168.123.105=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.077185+0000 mgr.a (mgr.14150) 31 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm01:192.168.123.101=a;vm05:192.168.123.105=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: cephadm 2026-03-10T12:33:28.078210+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Saving service mon spec with placement vm01:192.168.123.101=a;vm05:192.168.123.105=b;count:2 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: cephadm 2026-03-10T12:33:28.078210+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Saving service mon spec with placement vm01:192.168.123.101=a;vm05:192.168.123.105=b;count:2 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.081014+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.081014+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.081869+0000 mon.a (mon.0) 161 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.081869+0000 mon.a (mon.0) 161 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.083399+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.083399+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.083885+0000 mon.a (mon.0) 163 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.083885+0000 mon.a (mon.0) 163 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.086504+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.086504+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.087936+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.087936+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.088576+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: audit 2026-03-10T12:33:28.088576+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: cephadm 2026-03-10T12:33:28.089263+0000 mgr.a (mgr.14150) 33 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:33:29.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:29 vm01 bash[20716]: cephadm 2026-03-10T12:33:28.089263+0000 mgr.a (mgr.14150) 33 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:33:29.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:29 vm05 systemd[1]: Started Ceph mon.b for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:33:29.978 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:33:29.978 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":1,"fsid":"f864ef90-1c7c-11f1-946a-0b343fff599a","modified":"2026-03-10T12:32:35.364821Z","created":"2026-03-10T12:32:35.364821Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T12:33:29.978 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 1 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 0 pidfile_write: ignore empty --pid-file 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 0 load: jerasure load: lrc 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Git sha 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: DB SUMMARY 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: DB Session ID: W50K6W9HURIGSNF8MHD1 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 0, files: 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000004.log size: 511 ; 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.env: 0x56144e2badc0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.info_log: 0x56145fedc700 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.statistics: (nil) 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.use_fsync: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.db_log_dir: 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.wal_dir: 2026-03-10T12:33:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.write_buffer_manager: 0x56145fee1900 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.unordered_write: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.row_cache: None 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.wal_filter: None 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.wal_compression: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T12:33:30.176 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_open_files: -1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Compression algorithms supported: 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: kZSTD supported: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.merge_operator: 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_filter: None 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x56145fedc640) 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cache_index_and_filter_blocks: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: pin_top_level_index_and_filter: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: index_type: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: data_block_index_type: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: index_shortening: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: checksum: 4 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: no_block_cache: 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: block_cache: 0x56145ff03350 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: block_cache_name: BinnedLRUCache 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: block_cache_options: 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: capacity : 536870912 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: num_shard_bits : 4 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: strict_capacity_limit : 0 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: high_pri_pool_ratio: 0.000 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: block_cache_compressed: (nil) 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: persistent_cache: (nil) 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: block_size: 4096 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: block_size_deviation: 10 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: block_restart_interval: 16 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: index_block_restart_interval: 1 2026-03-10T12:33:30.177 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: metadata_block_size: 4096 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: partition_filters: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: use_delta_encoding: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: filter_policy: bloomfilter 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: whole_key_filtering: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: verify_compression: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: read_amp_bytes_per_bit: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: format_version: 5 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: enable_index_compression: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: block_align: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: max_auto_readahead_size: 262144 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: prepopulate_block_cache: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: initial_auto_readahead_size: 8192 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: num_file_reads_for_auto_readahead: 2 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression: NoCompression 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.num_levels: 7 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T12:33:30.178 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.ttl: 2592000 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.082+0000 7fb348c15d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.086+0000 7fb348c15d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.086+0000 7fb348c15d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.086+0000 7fb348c15d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 6bd5b8b6-8815-4e14-b03d-5213c86c0db1 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.086+0000 7fb348c15d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146010087820, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.086+0000 7fb348c15d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.086+0000 7fb348c15d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146010089900, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773146010, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "6bd5b8b6-8815-4e14-b03d-5213c86c0db1", "db_session_id": "W50K6W9HURIGSNF8MHD1", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.086+0000 7fb348c15d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146010089955, "job": 1, "event": "recovery_finished"} 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.086+0000 7fb348c15d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb348c15d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb348c15d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x56145ff04e00 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb348c15d80 4 rocksdb: DB pointer 0x56146001a000 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb348c15d80 0 mon.b does not exist in monmap, will attempt to join an existing cluster 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb33e9df640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb33e9df640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: ** DB Stats ** 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T12:33:30.179 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: ** Compaction Stats [default] ** 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.8 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.8 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.8 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: ** Compaction Stats [default] ** 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.8 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: AddFile(Keys): cumulative 0, interval 0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Cumulative compaction: 0.00 GB write, 0.10 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Interval compaction: 0.00 GB write, 0.10 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Block cache BinnedLRUCache@0x56145ff03350#7 capacity: 512.00 MB usage: 0.22 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 9e-06 secs_since: 0 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: Block cache entry stats(count,size,portion): FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: ** File Read Latency Histogram By Level [default] ** 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb348c15d80 0 using public_addr v2:192.168.123.105:0/0 -> [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb348c15d80 0 starting mon.b rank -1 at public addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] at bind addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:30.180 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.098+0000 7fb348c15d80 1 mon.b@-1(???) e0 preinit fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 0 mon.b@-1(synchronizing).mds e1 new map 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 0 mon.b@-1(synchronizing).mds e1 print_map 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: e1 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: btime 2026-03-10T12:32:36:745933+0000 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: legacy client fscid: -1 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: No filesystems configured 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 1 mon.b@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 1 mon.b@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 1 mon.b@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 1 mon.b@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 1 mon.b@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 1 mon.b@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 0 mon.b@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.122+0000 7fb3419e5640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:36.746413+0000 mon.a (mon.0) 0 : cluster [INF] mkfs f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:36.746413+0000 mon.a (mon.0) 0 : cluster [INF] mkfs f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:36.741242+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:36.741242+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985791+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985791+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985834+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985834+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985840+0000 mon.a (mon.0) 3 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985840+0000 mon.a (mon.0) 3 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985843+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T12:32:35.364821+0000 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985843+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T12:32:35.364821+0000 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985851+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985851+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985855+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985855+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985860+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985860+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985863+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.985863+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.986107+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.986107+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.986121+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.986121+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.986638+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:37.986638+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:38.032213+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.101:0/586937715' entity='client.admin' 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:38.032213+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.101:0/586937715' entity='client.admin' 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:38.654260+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.101:0/4165219441' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:38.654260+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.101:0/4165219441' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:40.886421+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.101:0/670267606' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:40.886421+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.101:0/670267606' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:41.899560+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:41.899560+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T12:33:30.181 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:41.904872+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00542696s) 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:41.904872+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00542696s) 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.907537+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.907537+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.908004+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.908004+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.908414+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.908414+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.908904+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.908904+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.910052+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.910052+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:41.916614+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:41.916614+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.929855+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.929855+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.931607+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.931607+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.935085+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.935085+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.936540+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.936540+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.944600+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:41.944600+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.101:0/3907530188' entity='mgr.a' 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:42.909841+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0104s) 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:42.909841+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0104s) 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:43.224964+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.101:0/3145018730' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:43.224964+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.101:0/3145018730' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:43.496151+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.101:0/4287141444' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:43.496151+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.101:0/4287141444' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:43.498855+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.101:0/4287141444' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:43.498855+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.101:0/4287141444' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:43.799213+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.101:0/2672616149' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:43.799213+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.101:0/2672616149' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:44.499484+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.101:0/2672616149' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:44.499484+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.101:0/2672616149' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:44.502305+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:44.502305+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:44.876320+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.101:0/1608840306' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:44.876320+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.101:0/1608840306' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.865370+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.865370+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.865781+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.865781+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.869569+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.869569+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.869660+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00397717s) 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.869660+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00397717s) 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.872484+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.872484+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.872844+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.872844+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.874161+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.874161+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.874537+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.874537+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.874886+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:33:30.182 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.874886+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.881414+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:47.881414+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.890921+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.890921+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.894237+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.894237+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.896388+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.896388+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.908480+0000 mon.a (mon.0) 49 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.908480+0000 mon.a (mon.0) 49 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.910074+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.910074+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.910409+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.910409+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.916967+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:47.916967+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:47.888805+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:47.888805+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:48.384448+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:48.384448+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:48.387537+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:48.387537+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:48.875057+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00936s) 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:48.875057+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00936s) 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.208682+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.208682+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.213534+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.213534+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:48.875909+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:48.875909+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:48.880956+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:48.880956+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.205182+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.205182+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.242908+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Bus STARTING 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.242908+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Bus STARTING 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.359671+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.359671+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.360163+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Client ('192.168.123.101', 43636) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.360163+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Client ('192.168.123.101', 43636) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.460855+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.460855+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.460894+0000 mgr.a (mgr.14118) 9 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Bus STARTED 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.460894+0000 mgr.a (mgr.14118) 9 : cephadm [INF] [10/Mar/2026:12:32:49] ENGINE Bus STARTED 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.461711+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.461711+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.468411+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.468411+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.183 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.871095+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.871095+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.873857+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.873857+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.849178+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:49.849178+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.849411+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:49.849411+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:50.271865+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:50.271865+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:50.389118+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:50.389118+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:50.538622+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:50.538622+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:51.088722+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm01 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:51.088722+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm01 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.335859+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.335859+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.338872+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.338872+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:52.336178+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm01 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:52.336178+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm01 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.667638+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.667638+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.941035+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.941035+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:53.205811+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.101:0/2874561222' entity='client.admin' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:53.205811+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.101:0/2874561222' entity='client.admin' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.663571+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.663571+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:52.664464+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:52.664464+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.937111+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:52.937111+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:52.937809+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:52.937809+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:53.481442+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.101:0/3519582229' entity='client.admin' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:53.481442+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.101:0/3519582229' entity='client.admin' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:53.812170+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:53.812170+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:53.831814+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.101:0/2164443391' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:53.831814+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.101:0/2164443391' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:54.113979+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:54.113979+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.101:0/2655754207' entity='mgr.a' 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:54.812436+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.101:0/2164443391' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:54.812436+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.101:0/2164443391' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:54.818334+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:54.818334+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:55.211171+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.101:0/2437500333' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:55.211171+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.101:0/2437500333' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.023943+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.023943+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.024338+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.024338+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.028461+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.028461+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.028600+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00435932s) 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.028600+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00435932s) 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.032632+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.032632+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:30.184 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.032978+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.032978+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.033497+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.033497+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.033879+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.033879+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.034230+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.034230+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.039219+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:58.039219+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.050802+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.050802+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.063683+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.063683+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.076339+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.076339+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.088275+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:58.088275+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.711027+0000 mgr.a (mgr.14150) 1 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Bus STARTING 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.711027+0000 mgr.a (mgr.14150) 1 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Bus STARTING 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.812622+0000 mgr.a (mgr.14150) 2 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.812622+0000 mgr.a (mgr.14150) 2 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:59.032279+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00804s) 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:32:59.032279+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00804s) 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.921725+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.921725+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.921769+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Bus STARTED 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.921769+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Bus STARTED 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.922173+0000 mgr.a (mgr.14150) 5 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Client ('192.168.123.101', 54550) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:32:58.922173+0000 mgr.a (mgr.14150) 5 : cephadm [INF] [10/Mar/2026:12:32:58] ENGINE Client ('192.168.123.101', 54550) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.033545+0000 mgr.a (mgr.14150) 6 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.033545+0000 mgr.a (mgr.14150) 6 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.037630+0000 mgr.a (mgr.14150) 7 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.037630+0000 mgr.a (mgr.14150) 7 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.320182+0000 mgr.a (mgr.14150) 8 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.320182+0000 mgr.a (mgr.14150) 8 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.364047+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.364047+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.367533+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.367533+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.628654+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.628654+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.783799+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:32:59.783799+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:00.099772+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.101:0/4077019482' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:00.099772+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.101:0/4077019482' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:00.433950+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.101:0/1610423079' entity='client.admin' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:00.433950+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.101:0/1610423079' entity='client.admin' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:00.786608+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:00.786608+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.052130+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.052130+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.052676+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.052676+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.055480+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.055480+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.056423+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.056423+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]: dispatch 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.057352+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]': finished 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:03.057352+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]': finished 2026-03-10T12:33:30.185 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:03.604092+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm01 on vm01 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:03.604092+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm01 on vm01 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:04.106580+0000 mon.a (mon.0) 100 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:04.106580+0000 mon.a (mon.0) 100 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.305899+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.305899+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.309992+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.309992+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.312984+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.312984+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.320691+0000 mon.a (mon.0) 104 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.320691+0000 mon.a (mon.0) 104 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.324338+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.324338+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.326587+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.326587+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.328762+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.328762+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.330890+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.330890+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.333743+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.333743+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.340509+0000 mon.a (mon.0) 110 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.340509+0000 mon.a (mon.0) 110 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.344156+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.344156+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.616732+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:04.616732+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:05.159316+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.101:0/2715616957' entity='client.admin' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:05.159316+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.101:0/2715616957' entity='client.admin' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:05.165650+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:05.165650+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:05.169704+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:05.169704+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:08.060690+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:08.060690+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.178781+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.178781+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.181662+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.181662+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.201990+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.201990+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.204886+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.204886+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.205707+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.205707+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.206748+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.206748+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.207089+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.207089+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:09.207561+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm01:/etc/ceph/ceph.conf 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:09.207561+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm01:/etc/ceph/ceph.conf 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:09.250601+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:30.186 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:09.250601+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:09.294165+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:09.294165+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:09.326858+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.client.admin.keyring 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:09.326858+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.client.admin.keyring 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.364569+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.364569+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.367489+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:09.367489+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:14.291682+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:14.291682+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:14.294448+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:14.294448+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:15.000834+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:15.000834+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:15.540621+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm05 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:15.540621+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm05 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:16.799193+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:16.799193+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:16.799687+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm05 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:16.799687+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm05 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:16.799958+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:16.799958+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:17.089955+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:17.089955+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:18.035857+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:18.035857+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.376240+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.376240+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.377999+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.377999+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.378406+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.378406+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.378981+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.378981+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.379383+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.379383+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.379940+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.379940+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.407710+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.407710+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.433978+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.433978+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.460239+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.client.admin.keyring 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.460239+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.client.admin.keyring 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.490650+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.187 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.490650+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.493028+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.493028+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.493945+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm05", "caps": []}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.493945+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm05", "caps": []}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.495000+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm05", "caps": []}]': finished 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:18.495000+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm05", "caps": []}]': finished 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.948412+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm05 on vm05 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:18.948412+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm05 on vm05 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.610963+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.610963+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.613484+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.613484+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.615816+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.615816+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.625000+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.625000+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.625982+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.625982+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.626315+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.626315+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.629103+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.629103+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.637560+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.637560+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.638297+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.638297+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.638681+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.638681+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.640876+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.640876+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.896139+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:19.896139+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:20.036015+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:20.036015+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:20.957904+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:20.957904+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:20.960135+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:20.960135+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:21.725771+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:21.725771+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:22.036295+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:22.036295+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:23.063503+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:23.063503+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:24.036499+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:24.036499+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:25.773202+0000 mon.a (mon.0) 154 : audit [INF] from='client.? 192.168.123.101:0/378617481' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:25.773202+0000 mon.a (mon.0) 154 : audit [INF] from='client.? 192.168.123.101:0/378617481' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:26.036682+0000 mgr.a (mgr.14150) 29 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:26.036682+0000 mgr.a (mgr.14150) 29 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:26.627971+0000 mon.a (mon.0) 155 : audit [INF] from='client.? 192.168.123.101:0/378617481' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:26.627971+0000 mon.a (mon.0) 155 : audit [INF] from='client.? 192.168.123.101:0/378617481' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:26.632410+0000 mon.a (mon.0) 156 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:26.632410+0000 mon.a (mon.0) 156 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:27.598342+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:27.598342+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.188 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:27.602006+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:27.602006+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:27.605250+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:27.605250+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:28.036934+0000 mgr.a (mgr.14150) 30 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cluster 2026-03-10T12:33:28.036934+0000 mgr.a (mgr.14150) 30 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.077185+0000 mgr.a (mgr.14150) 31 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm01:192.168.123.101=a;vm05:192.168.123.105=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.077185+0000 mgr.a (mgr.14150) 31 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm01:192.168.123.101=a;vm05:192.168.123.105=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:28.078210+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Saving service mon spec with placement vm01:192.168.123.101=a;vm05:192.168.123.105=b;count:2 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:28.078210+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Saving service mon spec with placement vm01:192.168.123.101=a;vm05:192.168.123.105=b;count:2 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.081014+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.081014+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.081869+0000 mon.a (mon.0) 161 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.081869+0000 mon.a (mon.0) 161 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.083399+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.083399+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.083885+0000 mon.a (mon.0) 163 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.083885+0000 mon.a (mon.0) 163 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.086504+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.086504+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.087936+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.087936+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.088576+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: audit 2026-03-10T12:33:28.088576+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:28.089263+0000 mgr.a (mgr.14150) 33 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: cephadm 2026-03-10T12:33:28.089263+0000 mgr.a (mgr.14150) 33 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:33:30.189 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:30 vm05 bash[20951]: debug 2026-03-10T12:33:30.130+0000 7fb3419e5640 1 mon.b@-1(synchronizing).paxosservice(auth 1..5) refresh upgraded, format 0 -> 3 2026-03-10T12:33:31.107 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T12:33:31.107 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph mon dump -f json 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:30.037133+0000 mgr.a (mgr.14150) 37 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:30.037133+0000 mgr.a (mgr.14150) 37 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:30.143436+0000 mon.a (mon.0) 179 : cluster [INF] mon.a calling monitor election 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:30.143436+0000 mon.a (mon.0) 179 : cluster [INF] mon.a calling monitor election 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:30.144717+0000 mon.a (mon.0) 180 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:30.144717+0000 mon.a (mon.0) 180 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:30.144775+0000 mon.a (mon.0) 181 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:30.144775+0000 mon.a (mon.0) 181 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.620 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:31.137167+0000 mon.a (mon.0) 182 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:31.137167+0000 mon.a (mon.0) 182 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:32.037358+0000 mgr.a (mgr.14150) 38 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:32.037358+0000 mgr.a (mgr.14150) 38 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:32.137336+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:32.137336+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:32.141029+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:32.141029+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:33.137950+0000 mon.a (mon.0) 184 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:33.137950+0000 mon.a (mon.0) 184 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:34.037617+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:34.037617+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:34.137753+0000 mon.a (mon.0) 185 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:34.137753+0000 mon.a (mon.0) 185 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.137575+0000 mon.a (mon.0) 186 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.137575+0000 mon.a (mon.0) 186 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.165691+0000 mon.a (mon.0) 187 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.165691+0000 mon.a (mon.0) 187 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169341+0000 mon.a (mon.0) 188 : cluster [DBG] monmap epoch 2 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169341+0000 mon.a (mon.0) 188 : cluster [DBG] monmap epoch 2 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169370+0000 mon.a (mon.0) 189 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169370+0000 mon.a (mon.0) 189 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169381+0000 mon.a (mon.0) 190 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169381+0000 mon.a (mon.0) 190 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169392+0000 mon.a (mon.0) 191 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169392+0000 mon.a (mon.0) 191 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169403+0000 mon.a (mon.0) 192 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169403+0000 mon.a (mon.0) 192 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169414+0000 mon.a (mon.0) 193 : cluster [DBG] election_strategy: 1 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169414+0000 mon.a (mon.0) 193 : cluster [DBG] election_strategy: 1 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169427+0000 mon.a (mon.0) 194 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169427+0000 mon.a (mon.0) 194 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169438+0000 mon.a (mon.0) 195 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169438+0000 mon.a (mon.0) 195 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169877+0000 mon.a (mon.0) 196 : cluster [DBG] fsmap 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169877+0000 mon.a (mon.0) 196 : cluster [DBG] fsmap 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169903+0000 mon.a (mon.0) 197 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.169903+0000 mon.a (mon.0) 197 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.170078+0000 mon.a (mon.0) 198 : cluster [DBG] mgrmap e12: a(active, since 37s) 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.170078+0000 mon.a (mon.0) 198 : cluster [DBG] mgrmap e12: a(active, since 37s) 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.170185+0000 mon.a (mon.0) 199 : cluster [INF] overall HEALTH_OK 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: cluster 2026-03-10T12:33:35.170185+0000 mon.a (mon.0) 199 : cluster [INF] overall HEALTH_OK 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.173641+0000 mon.a (mon.0) 200 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.173641+0000 mon.a (mon.0) 200 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.175731+0000 mon.a (mon.0) 201 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.175731+0000 mon.a (mon.0) 201 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.179625+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.179625+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.183783+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.183783+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.185927+0000 mon.a (mon.0) 204 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.185927+0000 mon.a (mon.0) 204 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.186544+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.186544+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.186987+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:35.621 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:35 vm05 bash[20951]: audit 2026-03-10T12:33:35.186987+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:35.644 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.b/config 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:30.037133+0000 mgr.a (mgr.14150) 37 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:30.037133+0000 mgr.a (mgr.14150) 37 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:30.143436+0000 mon.a (mon.0) 179 : cluster [INF] mon.a calling monitor election 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:30.143436+0000 mon.a (mon.0) 179 : cluster [INF] mon.a calling monitor election 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:30.144717+0000 mon.a (mon.0) 180 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:30.144717+0000 mon.a (mon.0) 180 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:30.144775+0000 mon.a (mon.0) 181 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:30.144775+0000 mon.a (mon.0) 181 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:31.137167+0000 mon.a (mon.0) 182 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:31.137167+0000 mon.a (mon.0) 182 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:32.037358+0000 mgr.a (mgr.14150) 38 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:32.037358+0000 mgr.a (mgr.14150) 38 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:32.137336+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:32.137336+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:32.141029+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:32.141029+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:33.137950+0000 mon.a (mon.0) 184 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:33.137950+0000 mon.a (mon.0) 184 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:34.037617+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:34.037617+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:34.137753+0000 mon.a (mon.0) 185 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:34.137753+0000 mon.a (mon.0) 185 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.137575+0000 mon.a (mon.0) 186 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.137575+0000 mon.a (mon.0) 186 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.165691+0000 mon.a (mon.0) 187 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.165691+0000 mon.a (mon.0) 187 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169341+0000 mon.a (mon.0) 188 : cluster [DBG] monmap epoch 2 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169341+0000 mon.a (mon.0) 188 : cluster [DBG] monmap epoch 2 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169370+0000 mon.a (mon.0) 189 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169370+0000 mon.a (mon.0) 189 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169381+0000 mon.a (mon.0) 190 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169381+0000 mon.a (mon.0) 190 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169392+0000 mon.a (mon.0) 191 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169392+0000 mon.a (mon.0) 191 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169403+0000 mon.a (mon.0) 192 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169403+0000 mon.a (mon.0) 192 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169414+0000 mon.a (mon.0) 193 : cluster [DBG] election_strategy: 1 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169414+0000 mon.a (mon.0) 193 : cluster [DBG] election_strategy: 1 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169427+0000 mon.a (mon.0) 194 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169427+0000 mon.a (mon.0) 194 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169438+0000 mon.a (mon.0) 195 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169438+0000 mon.a (mon.0) 195 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169877+0000 mon.a (mon.0) 196 : cluster [DBG] fsmap 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169877+0000 mon.a (mon.0) 196 : cluster [DBG] fsmap 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169903+0000 mon.a (mon.0) 197 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.169903+0000 mon.a (mon.0) 197 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.170078+0000 mon.a (mon.0) 198 : cluster [DBG] mgrmap e12: a(active, since 37s) 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.170078+0000 mon.a (mon.0) 198 : cluster [DBG] mgrmap e12: a(active, since 37s) 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.170185+0000 mon.a (mon.0) 199 : cluster [INF] overall HEALTH_OK 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: cluster 2026-03-10T12:33:35.170185+0000 mon.a (mon.0) 199 : cluster [INF] overall HEALTH_OK 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.173641+0000 mon.a (mon.0) 200 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.173641+0000 mon.a (mon.0) 200 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.175731+0000 mon.a (mon.0) 201 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.175731+0000 mon.a (mon.0) 201 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.179625+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.179625+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.183783+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.183783+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.185927+0000 mon.a (mon.0) 204 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.185927+0000 mon.a (mon.0) 204 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.186544+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.186544+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.186987+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:35.701 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:35 vm01 bash[20716]: audit 2026-03-10T12:33:35.186987+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.039 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:33:36.039 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":2,"fsid":"f864ef90-1c7c-11f1-946a-0b343fff599a","modified":"2026-03-10T12:33:30.138463Z","created":"2026-03-10T12:32:35.364821Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:3300","nonce":0},{"type":"v1","addr":"192.168.123.105:6789","nonce":0}]},"addr":"192.168.123.105:6789/0","public_addr":"192.168.123.105:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-10T12:33:36.039 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 2 2026-03-10T12:33:36.106 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-10T12:33:36.106 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph config generate-minimal-conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.184774+0000 mgr.a (mgr.14150) 40 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.184774+0000 mgr.a (mgr.14150) 40 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.187536+0000 mgr.a (mgr.14150) 41 : cephadm [INF] Reconfiguring daemon mon.b on vm05 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.187536+0000 mgr.a (mgr.14150) 41 : cephadm [INF] Reconfiguring daemon mon.b on vm05 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.259910+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.259910+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.814187+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.814187+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.818748+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.818748+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.819694+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.819694+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.820854+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.820854+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.821255+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.821255+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.821773+0000 mgr.a (mgr.14150) 42 : cephadm [INF] Updating vm01:/etc/ceph/ceph.conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.821773+0000 mgr.a (mgr.14150) 42 : cephadm [INF] Updating vm01:/etc/ceph/ceph.conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.821874+0000 mgr.a (mgr.14150) 43 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.821874+0000 mgr.a (mgr.14150) 43 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.866402+0000 mgr.a (mgr.14150) 44 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.866402+0000 mgr.a (mgr.14150) 44 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.867509+0000 mgr.a (mgr.14150) 45 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.867509+0000 mgr.a (mgr.14150) 45 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.915977+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.915977+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.919361+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.919361+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.922588+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.922588+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.925775+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.925775+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.929388+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.929388+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.941153+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.941153+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.941697+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.941697+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.942206+0000 mon.a (mon.0) 220 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:35.942206+0000 mon.a (mon.0) 220 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:36.038042+0000 mon.a (mon.0) 221 : audit [DBG] from='client.? 192.168.123.105:0/528992565' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:36.038042+0000 mon.a (mon.0) 221 : audit [DBG] from='client.? 192.168.123.105:0/528992565' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:36.138402+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:36.426 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:36 vm05 bash[20951]: audit 2026-03-10T12:33:36.138402+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.184774+0000 mgr.a (mgr.14150) 40 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.184774+0000 mgr.a (mgr.14150) 40 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.187536+0000 mgr.a (mgr.14150) 41 : cephadm [INF] Reconfiguring daemon mon.b on vm05 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.187536+0000 mgr.a (mgr.14150) 41 : cephadm [INF] Reconfiguring daemon mon.b on vm05 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.259910+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.259910+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.814187+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.814187+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.818748+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.818748+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.819694+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.819694+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.820854+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.820854+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.821255+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.821255+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.821773+0000 mgr.a (mgr.14150) 42 : cephadm [INF] Updating vm01:/etc/ceph/ceph.conf 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.821773+0000 mgr.a (mgr.14150) 42 : cephadm [INF] Updating vm01:/etc/ceph/ceph.conf 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.821874+0000 mgr.a (mgr.14150) 43 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.821874+0000 mgr.a (mgr.14150) 43 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.866402+0000 mgr.a (mgr.14150) 44 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.866402+0000 mgr.a (mgr.14150) 44 : cephadm [INF] Updating vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:36.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.867509+0000 mgr.a (mgr.14150) 45 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.867509+0000 mgr.a (mgr.14150) 45 : cephadm [INF] Updating vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/config/ceph.conf 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.915977+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.915977+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.919361+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.919361+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.922588+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.922588+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.925775+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.925775+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.929388+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.929388+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.941153+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.941153+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.941697+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.941697+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.942206+0000 mon.a (mon.0) 220 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:35.942206+0000 mon.a (mon.0) 220 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:36.038042+0000 mon.a (mon.0) 221 : audit [DBG] from='client.? 192.168.123.105:0/528992565' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:36.038042+0000 mon.a (mon.0) 221 : audit [DBG] from='client.? 192.168.123.105:0/528992565' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:36.138402+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:36.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:36 vm01 bash[20716]: audit 2026-03-10T12:33:36.138402+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:33:37.449 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20987]: debug 2026-03-10T12:33:37.132+0000 7f5439b90640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-10T12:33:37.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.929963+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:37.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.929963+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.940805+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.940805+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.942789+0000 mgr.a (mgr.14150) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm01 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: cephadm 2026-03-10T12:33:35.942789+0000 mgr.a (mgr.14150) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm01 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: cluster 2026-03-10T12:33:36.037851+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: cluster 2026-03-10T12:33:36.037851+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: audit 2026-03-10T12:33:36.348934+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: audit 2026-03-10T12:33:36.348934+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: audit 2026-03-10T12:33:36.352937+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:37.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:37 vm01 bash[20716]: audit 2026-03-10T12:33:36.352937+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:37.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.929963+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.929963+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.940805+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.940805+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.942789+0000 mgr.a (mgr.14150) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm01 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: cephadm 2026-03-10T12:33:35.942789+0000 mgr.a (mgr.14150) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm01 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: cluster 2026-03-10T12:33:36.037851+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: cluster 2026-03-10T12:33:36.037851+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: audit 2026-03-10T12:33:36.348934+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: audit 2026-03-10T12:33:36.348934+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: audit 2026-03-10T12:33:36.352937+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:37.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:37 vm05 bash[20951]: audit 2026-03-10T12:33:36.352937+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.898582+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.898582+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.902263+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.902263+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.903204+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.903204+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.904873+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.904873+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.905799+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.905799+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.905882+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.905882+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.909787+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: audit 2026-03-10T12:33:37.909787+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: cluster 2026-03-10T12:33:38.038053+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:39.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:38 vm05 bash[20951]: cluster 2026-03-10T12:33:38.038053+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.898582+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.898582+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.902263+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.902263+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.903204+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.903204+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.904873+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.904873+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.905799+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.905799+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.905882+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.905882+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.909787+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: audit 2026-03-10T12:33:37.909787+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: cluster 2026-03-10T12:33:38.038053+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:39.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:38 vm01 bash[20716]: cluster 2026-03-10T12:33:38.038053+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:40.741 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:33:40.987 INFO:teuthology.orchestra.run.vm01.stdout:# minimal ceph.conf for f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:40.987 INFO:teuthology.orchestra.run.vm01.stdout:[global] 2026-03-10T12:33:40.987 INFO:teuthology.orchestra.run.vm01.stdout: fsid = f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:33:40.987 INFO:teuthology.orchestra.run.vm01.stdout: mon_host = [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] 2026-03-10T12:33:41.531 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-10T12:33:41.531 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:33:41.531 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T12:33:41.538 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:33:41.538 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:41.587 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:33:41.587 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T12:33:41.594 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:33:41.594 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:33:41.642 INFO:tasks.cephadm:Adding mgr.a on vm01 2026-03-10T12:33:41.642 INFO:tasks.cephadm:Adding mgr.b on vm05 2026-03-10T12:33:41.642 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch apply mgr '2;vm01=a;vm05=b' 2026-03-10T12:33:41.695 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:41 vm01 bash[20716]: cluster 2026-03-10T12:33:40.038256+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:41.695 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:41 vm01 bash[20716]: cluster 2026-03-10T12:33:40.038256+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:41.695 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:41 vm01 bash[20716]: audit 2026-03-10T12:33:40.986196+0000 mon.a (mon.0) 232 : audit [DBG] from='client.? 192.168.123.101:0/1475709618' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:41.695 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:41 vm01 bash[20716]: audit 2026-03-10T12:33:40.986196+0000 mon.a (mon.0) 232 : audit [DBG] from='client.? 192.168.123.101:0/1475709618' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:41.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:41 vm05 bash[20951]: cluster 2026-03-10T12:33:40.038256+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:41.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:41 vm05 bash[20951]: cluster 2026-03-10T12:33:40.038256+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:41.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:41 vm05 bash[20951]: audit 2026-03-10T12:33:40.986196+0000 mon.a (mon.0) 232 : audit [DBG] from='client.? 192.168.123.101:0/1475709618' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:41.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:41 vm05 bash[20951]: audit 2026-03-10T12:33:40.986196+0000 mon.a (mon.0) 232 : audit [DBG] from='client.? 192.168.123.101:0/1475709618' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:43.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:43 vm01 bash[20716]: cluster 2026-03-10T12:33:42.038440+0000 mgr.a (mgr.14150) 52 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:43.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:43 vm01 bash[20716]: cluster 2026-03-10T12:33:42.038440+0000 mgr.a (mgr.14150) 52 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:43.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:43 vm05 bash[20951]: cluster 2026-03-10T12:33:42.038440+0000 mgr.a (mgr.14150) 52 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:43.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:43 vm05 bash[20951]: cluster 2026-03-10T12:33:42.038440+0000 mgr.a (mgr.14150) 52 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:45.290 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.b/config 2026-03-10T12:33:45.554 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled mgr update... 2026-03-10T12:33:45.568 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:45 vm05 bash[20951]: cluster 2026-03-10T12:33:44.038639+0000 mgr.a (mgr.14150) 53 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:45.568 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:45 vm05 bash[20951]: cluster 2026-03-10T12:33:44.038639+0000 mgr.a (mgr.14150) 53 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:45.635 DEBUG:teuthology.orchestra.run.vm05:mgr.b> sudo journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.b.service 2026-03-10T12:33:45.636 INFO:tasks.cephadm:Deploying OSDs... 2026-03-10T12:33:45.636 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T12:33:45.636 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T12:33:45.639 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:33:45.640 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-10T12:33:45.685 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-10T12:33:45.685 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-10T12:33:45.685 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-10T12:33:45.685 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-10T12:33:45.685 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-10T12:33:45.685 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T12:33:45.685 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T12:33:45.685 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-10T12:33:45.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:45 vm01 bash[20716]: cluster 2026-03-10T12:33:44.038639+0000 mgr.a (mgr.14150) 53 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:45.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:45 vm01 bash[20716]: cluster 2026-03-10T12:33:44.038639+0000 mgr.a (mgr.14150) 53 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:45.701 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-10T12:33:45.701 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T12:33:45.701 INFO:teuthology.orchestra.run.vm01.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-10T12:33:45.701 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:33:45.701 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 12:26:33.327047233 +0000 2026-03-10T12:33:45.701 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 12:26:32.435047233 +0000 2026-03-10T12:33:45.701 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 12:26:32.435047233 +0000 2026-03-10T12:33:45.701 INFO:teuthology.orchestra.run.vm01.stdout: Birth: - 2026-03-10T12:33:45.701 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T12:33:45.749 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T12:33:45.749 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T12:33:45.749 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000120485 s, 4.2 MB/s 2026-03-10T12:33:45.750 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T12:33:45.794 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-10T12:33:45.825 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:45 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:45.838 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-10T12:33:45.838 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T12:33:45.838 INFO:teuthology.orchestra.run.vm01.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-10T12:33:45.838 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:33:45.838 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 12:26:33.335047233 +0000 2026-03-10T12:33:45.838 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 12:26:32.411047233 +0000 2026-03-10T12:33:45.838 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 12:26:32.411047233 +0000 2026-03-10T12:33:45.838 INFO:teuthology.orchestra.run.vm01.stdout: Birth: - 2026-03-10T12:33:45.838 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T12:33:45.885 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T12:33:45.885 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T12:33:45.885 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000152176 s, 3.4 MB/s 2026-03-10T12:33:45.886 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T12:33:45.931 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-10T12:33:45.978 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-10T12:33:45.978 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T12:33:45.978 INFO:teuthology.orchestra.run.vm01.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-10T12:33:45.978 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:33:45.978 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 12:26:33.327047233 +0000 2026-03-10T12:33:45.978 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 12:26:32.431047233 +0000 2026-03-10T12:33:45.978 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 12:26:32.431047233 +0000 2026-03-10T12:33:45.978 INFO:teuthology.orchestra.run.vm01.stdout: Birth: - 2026-03-10T12:33:45.978 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T12:33:46.026 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T12:33:46.026 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T12:33:46.026 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000148878 s, 3.4 MB/s 2026-03-10T12:33:46.027 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T12:33:46.071 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-10T12:33:46.113 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-10T12:33:46.113 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T12:33:46.113 INFO:teuthology.orchestra.run.vm01.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-10T12:33:46.113 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:33:46.113 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 12:26:33.335047233 +0000 2026-03-10T12:33:46.113 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 12:26:32.415047233 +0000 2026-03-10T12:33:46.113 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 12:26:32.415047233 +0000 2026-03-10T12:33:46.113 INFO:teuthology.orchestra.run.vm01.stdout: Birth: - 2026-03-10T12:33:46.114 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T12:33:46.160 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T12:33:46.160 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T12:33:46.160 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000126257 s, 4.1 MB/s 2026-03-10T12:33:46.161 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T12:33:46.207 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T12:33:46.207 DEBUG:teuthology.orchestra.run.vm05:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T12:33:46.210 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:33:46.210 DEBUG:teuthology.orchestra.run.vm05:> ls /dev/[sv]d? 2026-03-10T12:33:46.254 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vda 2026-03-10T12:33:46.254 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdb 2026-03-10T12:33:46.254 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdc 2026-03-10T12:33:46.255 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdd 2026-03-10T12:33:46.255 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vde 2026-03-10T12:33:46.255 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T12:33:46.255 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T12:33:46.255 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdb 2026-03-10T12:33:46.303 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdb 2026-03-10T12:33:46.304 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T12:33:46.304 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-10T12:33:46.304 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:33:46.304 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-10 12:26:58.480113496 +0000 2026-03-10T12:33:46.304 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-10 12:26:57.540113496 +0000 2026-03-10T12:33:46.304 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-10 12:26:57.540113496 +0000 2026-03-10T12:33:46.304 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-10T12:33:46.304 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T12:33:46.353 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:46.354 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:46.354 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:46 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:46.354 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:46 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:46.354 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:46 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:46.354 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:46 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:33:46.355 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-10T12:33:46.355 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-10T12:33:46.355 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000179095 s, 2.9 MB/s 2026-03-10T12:33:46.356 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T12:33:46.414 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdc 2026-03-10T12:33:46.458 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdc 2026-03-10T12:33:46.458 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T12:33:46.458 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-10T12:33:46.458 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:33:46.458 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-10 12:26:58.488113496 +0000 2026-03-10T12:33:46.458 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-10 12:26:57.568113496 +0000 2026-03-10T12:33:46.458 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-10 12:26:57.568113496 +0000 2026-03-10T12:33:46.458 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-10T12:33:46.458 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T12:33:46.509 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-10T12:33:46.509 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-10T12:33:46.509 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000106309 s, 4.8 MB/s 2026-03-10T12:33:46.511 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T12:33:46.558 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdd 2026-03-10T12:33:46.601 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdd 2026-03-10T12:33:46.601 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T12:33:46.601 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-10T12:33:46.601 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:33:46.601 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-10 12:26:58.476113496 +0000 2026-03-10T12:33:46.601 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-10 12:26:57.576113496 +0000 2026-03-10T12:33:46.601 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-10 12:26:57.576113496 +0000 2026-03-10T12:33:46.601 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-10T12:33:46.601 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T12:33:46.623 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:46 vm05 systemd[1]: Started Ceph mgr.b for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[22215]: debug 2026-03-10T12:33:46.586+0000 7f0d5f18d140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.548269+0000 mgr.a (mgr.14150) 54 : audit [DBG] from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm01=a;vm05=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.548269+0000 mgr.a (mgr.14150) 54 : audit [DBG] from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm01=a;vm05=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: cephadm 2026-03-10T12:33:45.549310+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Saving service mgr spec with placement vm01=a;vm05=b;count:2 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: cephadm 2026-03-10T12:33:45.549310+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Saving service mgr spec with placement vm01=a;vm05=b;count:2 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.552853+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.552853+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.553355+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.553355+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.554364+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.554364+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.554758+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.554758+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.557687+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.557687+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.558521+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.558521+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.559995+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.559995+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.561030+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.561030+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.561519+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:45.561519+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: cephadm 2026-03-10T12:33:45.562041+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Deploying daemon mgr.b on vm05 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: cephadm 2026-03-10T12:33:45.562041+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Deploying daemon mgr.b on vm05 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.412432+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.412432+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.415128+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.415128+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.417705+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.624 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.417705+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.420673+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.420673+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.435282+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.435282+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.436217+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.436217+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.436568+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.436568+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.440026+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.440026+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.449079+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.449079+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.449592+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.449592+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.449951+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.625 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[20951]: audit 2026-03-10T12:33:46.449951+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.630 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-10T12:33:46.630 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-10T12:33:46.630 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000127529 s, 4.0 MB/s 2026-03-10T12:33:46.630 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T12:33:46.680 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vde 2026-03-10T12:33:46.725 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vde 2026-03-10T12:33:46.725 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T12:33:46.725 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-10T12:33:46.725 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:33:46.725 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-10 12:26:58.488113496 +0000 2026-03-10T12:33:46.725 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-10 12:26:57.572113496 +0000 2026-03-10T12:33:46.725 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-10 12:26:57.572113496 +0000 2026-03-10T12:33:46.725 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-10T12:33:46.725 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T12:33:46.773 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-10T12:33:46.773 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-10T12:33:46.773 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000163496 s, 3.1 MB/s 2026-03-10T12:33:46.774 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T12:33:46.824 INFO:tasks.cephadm:Deploying osd.0 on vm01 with /dev/vde... 2026-03-10T12:33:46.824 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- lvm zap /dev/vde 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.548269+0000 mgr.a (mgr.14150) 54 : audit [DBG] from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm01=a;vm05=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.548269+0000 mgr.a (mgr.14150) 54 : audit [DBG] from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm01=a;vm05=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: cephadm 2026-03-10T12:33:45.549310+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Saving service mgr spec with placement vm01=a;vm05=b;count:2 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: cephadm 2026-03-10T12:33:45.549310+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Saving service mgr spec with placement vm01=a;vm05=b;count:2 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.552853+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.552853+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.553355+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.553355+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.554364+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.554364+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.554758+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.554758+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.557687+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.557687+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.558521+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.558521+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.559995+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.559995+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.561030+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.561030+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.561519+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:45.561519+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: cephadm 2026-03-10T12:33:45.562041+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Deploying daemon mgr.b on vm05 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: cephadm 2026-03-10T12:33:45.562041+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Deploying daemon mgr.b on vm05 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.412432+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.412432+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.415128+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.415128+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.417705+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.417705+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.420673+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.420673+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.435282+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.435282+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.436217+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.436217+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.436568+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:46.826 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.436568+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:46.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.440026+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.440026+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:46.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.449079+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:33:46.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.449079+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:33:46.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.449592+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:33:46.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.449592+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:33:46.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.449951+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.827 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:46 vm01 bash[20716]: audit 2026-03-10T12:33:46.449951+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:46.924 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[22215]: debug 2026-03-10T12:33:46.622+0000 7f0d5f18d140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T12:33:46.924 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:46 vm05 bash[22215]: debug 2026-03-10T12:33:46.742+0000 7f0d5f18d140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T12:33:47.424 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[22215]: debug 2026-03-10T12:33:47.078+0000 7f0d5f18d140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T12:33:47.827 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[22215]: debug 2026-03-10T12:33:47.598+0000 7f0d5f18d140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[22215]: debug 2026-03-10T12:33:47.690+0000 7f0d5f18d140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: cluster 2026-03-10T12:33:46.038808+0000 mgr.a (mgr.14150) 57 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: cluster 2026-03-10T12:33:46.038808+0000 mgr.a (mgr.14150) 57 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: cephadm 2026-03-10T12:33:46.440300+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: cephadm 2026-03-10T12:33:46.440300+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: cephadm 2026-03-10T12:33:46.448924+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: cephadm 2026-03-10T12:33:46.448924+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: cephadm 2026-03-10T12:33:46.450356+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mgr.a on vm01 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: cephadm 2026-03-10T12:33:46.450356+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mgr.a on vm01 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.912806+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.912806+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.922494+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.922494+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.931427+0000 mon.a (mon.0) 255 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.931427+0000 mon.a (mon.0) 255 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.939413+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.939413+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.940052+0000 mon.a (mon.0) 257 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.940052+0000 mon.a (mon.0) 257 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.948009+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.828 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[20951]: audit 2026-03-10T12:33:46.948009+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: cluster 2026-03-10T12:33:46.038808+0000 mgr.a (mgr.14150) 57 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: cluster 2026-03-10T12:33:46.038808+0000 mgr.a (mgr.14150) 57 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: cephadm 2026-03-10T12:33:46.440300+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: cephadm 2026-03-10T12:33:46.440300+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: cephadm 2026-03-10T12:33:46.448924+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: cephadm 2026-03-10T12:33:46.448924+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: cephadm 2026-03-10T12:33:46.450356+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mgr.a on vm01 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: cephadm 2026-03-10T12:33:46.450356+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mgr.a on vm01 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.912806+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.912806+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.922494+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.922494+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.931427+0000 mon.a (mon.0) 255 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.931427+0000 mon.a (mon.0) 255 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.939413+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.939413+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.940052+0000 mon.a (mon.0) 257 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.940052+0000 mon.a (mon.0) 257 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.948009+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:47.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:47 vm01 bash[20716]: audit 2026-03-10T12:33:46.948009+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.091 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[22215]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T12:33:48.091 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[22215]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T12:33:48.091 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[22215]: from numpy import show_config as show_numpy_config 2026-03-10T12:33:48.091 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[22215]: debug 2026-03-10T12:33:47.830+0000 7f0d5f18d140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T12:33:48.091 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:47 vm05 bash[22215]: debug 2026-03-10T12:33:47.970+0000 7f0d5f18d140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T12:33:48.091 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.010+0000 7f0d5f18d140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T12:33:48.091 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.046+0000 7f0d5f18d140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T12:33:48.354 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.086+0000 7f0d5f18d140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T12:33:48.354 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.138+0000 7f0d5f18d140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T12:33:48.924 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.670+0000 7f0d5f18d140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T12:33:48.924 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.710+0000 7f0d5f18d140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T12:33:48.924 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.746+0000 7f0d5f18d140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T12:33:48.924 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.886+0000 7f0d5f18d140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T12:33:48.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[20951]: cephadm 2026-03-10T12:33:46.949816+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:48.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[20951]: cephadm 2026-03-10T12:33:46.949816+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:48.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[20951]: audit 2026-03-10T12:33:47.667759+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[20951]: audit 2026-03-10T12:33:47.667759+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[20951]: audit 2026-03-10T12:33:47.670803+0000 mon.a (mon.0) 260 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[20951]: audit 2026-03-10T12:33:47.670803+0000 mon.a (mon.0) 260 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[20951]: audit 2026-03-10T12:33:47.676515+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[20951]: audit 2026-03-10T12:33:47.676515+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:48 vm01 bash[20716]: cephadm 2026-03-10T12:33:46.949816+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:48.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:48 vm01 bash[20716]: cephadm 2026-03-10T12:33:46.949816+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:33:48.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:48 vm01 bash[20716]: audit 2026-03-10T12:33:47.667759+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:48 vm01 bash[20716]: audit 2026-03-10T12:33:47.667759+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:48 vm01 bash[20716]: audit 2026-03-10T12:33:47.670803+0000 mon.a (mon.0) 260 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:48 vm01 bash[20716]: audit 2026-03-10T12:33:47.670803+0000 mon.a (mon.0) 260 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:48 vm01 bash[20716]: audit 2026-03-10T12:33:47.676515+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:48.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:48 vm01 bash[20716]: audit 2026-03-10T12:33:47.676515+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:49.226 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.926+0000 7f0d5f18d140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T12:33:49.227 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:48 vm05 bash[22215]: debug 2026-03-10T12:33:48.962+0000 7f0d5f18d140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T12:33:49.227 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[22215]: debug 2026-03-10T12:33:49.070+0000 7f0d5f18d140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:33:49.477 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[22215]: debug 2026-03-10T12:33:49.222+0000 7f0d5f18d140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T12:33:49.477 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[22215]: debug 2026-03-10T12:33:49.398+0000 7f0d5f18d140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T12:33:49.477 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[22215]: debug 2026-03-10T12:33:49.434+0000 7f0d5f18d140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T12:33:49.776 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[20951]: cluster 2026-03-10T12:33:48.039031+0000 mgr.a (mgr.14150) 62 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:49.776 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[20951]: cluster 2026-03-10T12:33:48.039031+0000 mgr.a (mgr.14150) 62 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:49.776 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[22215]: debug 2026-03-10T12:33:49.474+0000 7f0d5f18d140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T12:33:49.776 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[22215]: debug 2026-03-10T12:33:49.618+0000 7f0d5f18d140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:33:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:49 vm01 bash[20716]: cluster 2026-03-10T12:33:48.039031+0000 mgr.a (mgr.14150) 62 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:49 vm01 bash[20716]: cluster 2026-03-10T12:33:48.039031+0000 mgr.a (mgr.14150) 62 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:50.174 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:33:49 vm05 bash[22215]: debug 2026-03-10T12:33:49.850+0000 7f0d5f18d140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T12:33:51.055 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.774772+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.774772+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: cluster 2026-03-10T12:33:49.857249+0000 mon.a (mon.0) 263 : cluster [DBG] Standby manager daemon b started 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: cluster 2026-03-10T12:33:49.857249+0000 mon.a (mon.0) 263 : cluster [DBG] Standby manager daemon b started 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.857546+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.857546+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.858129+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.858129+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.859199+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.859199+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.859712+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:49.859712+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:50.184345+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:51.072 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:50 vm01 bash[20716]: audit 2026-03-10T12:33:50.184345+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.774772+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.774772+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: cluster 2026-03-10T12:33:49.857249+0000 mon.a (mon.0) 263 : cluster [DBG] Standby manager daemon b started 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: cluster 2026-03-10T12:33:49.857249+0000 mon.a (mon.0) 263 : cluster [DBG] Standby manager daemon b started 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.857546+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.857546+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.858129+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.858129+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.859199+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.859199+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.859712+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:49.859712+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.105:0/2403354772' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T12:33:51.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:50.184345+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:51.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:50 vm05 bash[20951]: audit 2026-03-10T12:33:50.184345+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:51.903 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:33:51.918 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch daemon add osd vm01:/dev/vde 2026-03-10T12:33:52.070 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:51 vm01 bash[20716]: cluster 2026-03-10T12:33:50.039314+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:52.070 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:51 vm01 bash[20716]: cluster 2026-03-10T12:33:50.039314+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:52.070 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:51 vm01 bash[20716]: cluster 2026-03-10T12:33:50.793521+0000 mon.a (mon.0) 265 : cluster [DBG] mgrmap e13: a(active, since 52s), standbys: b 2026-03-10T12:33:52.070 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:51 vm01 bash[20716]: cluster 2026-03-10T12:33:50.793521+0000 mon.a (mon.0) 265 : cluster [DBG] mgrmap e13: a(active, since 52s), standbys: b 2026-03-10T12:33:52.070 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:51 vm01 bash[20716]: audit 2026-03-10T12:33:50.793648+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T12:33:52.071 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:51 vm01 bash[20716]: audit 2026-03-10T12:33:50.793648+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T12:33:52.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:51 vm05 bash[20951]: cluster 2026-03-10T12:33:50.039314+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:52.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:51 vm05 bash[20951]: cluster 2026-03-10T12:33:50.039314+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:52.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:51 vm05 bash[20951]: cluster 2026-03-10T12:33:50.793521+0000 mon.a (mon.0) 265 : cluster [DBG] mgrmap e13: a(active, since 52s), standbys: b 2026-03-10T12:33:52.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:51 vm05 bash[20951]: cluster 2026-03-10T12:33:50.793521+0000 mon.a (mon.0) 265 : cluster [DBG] mgrmap e13: a(active, since 52s), standbys: b 2026-03-10T12:33:52.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:51 vm05 bash[20951]: audit 2026-03-10T12:33:50.793648+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T12:33:52.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:51 vm05 bash[20951]: audit 2026-03-10T12:33:50.793648+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T12:33:54.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:53 vm05 bash[20951]: cluster 2026-03-10T12:33:52.039495+0000 mgr.a (mgr.14150) 64 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:54.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:53 vm05 bash[20951]: cluster 2026-03-10T12:33:52.039495+0000 mgr.a (mgr.14150) 64 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:54.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:53 vm01 bash[20716]: cluster 2026-03-10T12:33:52.039495+0000 mgr.a (mgr.14150) 64 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:54.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:53 vm01 bash[20716]: cluster 2026-03-10T12:33:52.039495+0000 mgr.a (mgr.14150) 64 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.973538+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.973538+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.976807+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.976807+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.978964+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.978964+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.980065+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.980065+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.980652+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.980652+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.981579+0000 mon.a (mon.0) 272 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.981579+0000 mon.a (mon.0) 272 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.984990+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: audit 2026-03-10T12:33:53.984990+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: cluster 2026-03-10T12:33:54.039644+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:55.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:54 vm05 bash[20951]: cluster 2026-03-10T12:33:54.039644+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:55.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.973538+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.973538+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.976807+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.976807+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.978964+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.978964+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.980065+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.980065+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.980652+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.980652+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.981579+0000 mon.a (mon.0) 272 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.981579+0000 mon.a (mon.0) 272 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.984990+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: audit 2026-03-10T12:33:53.984990+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: cluster 2026-03-10T12:33:54.039644+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:55.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:54 vm01 bash[20716]: cluster 2026-03-10T12:33:54.039644+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:56.527 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:33:57.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: cluster 2026-03-10T12:33:56.039800+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: cluster 2026-03-10T12:33:56.039800+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: audit 2026-03-10T12:33:56.882915+0000 mgr.a (mgr.14150) 67 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: audit 2026-03-10T12:33:56.882915+0000 mgr.a (mgr.14150) 67 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: audit 2026-03-10T12:33:56.884373+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: audit 2026-03-10T12:33:56.884373+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: audit 2026-03-10T12:33:56.886470+0000 mon.a (mon.0) 275 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: audit 2026-03-10T12:33:56.886470+0000 mon.a (mon.0) 275 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: audit 2026-03-10T12:33:56.887020+0000 mon.a (mon.0) 276 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:57.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:57 vm05 bash[20951]: audit 2026-03-10T12:33:56.887020+0000 mon.a (mon.0) 276 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:57.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: cluster 2026-03-10T12:33:56.039800+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:57.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: cluster 2026-03-10T12:33:56.039800+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:57.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: audit 2026-03-10T12:33:56.882915+0000 mgr.a (mgr.14150) 67 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:57.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: audit 2026-03-10T12:33:56.882915+0000 mgr.a (mgr.14150) 67 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:33:57.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: audit 2026-03-10T12:33:56.884373+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:33:57.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: audit 2026-03-10T12:33:56.884373+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:33:57.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: audit 2026-03-10T12:33:56.886470+0000 mon.a (mon.0) 275 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:33:57.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: audit 2026-03-10T12:33:56.886470+0000 mon.a (mon.0) 275 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:33:57.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: audit 2026-03-10T12:33:56.887020+0000 mon.a (mon.0) 276 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:57.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:57 vm01 bash[20716]: audit 2026-03-10T12:33:56.887020+0000 mon.a (mon.0) 276 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:33:59.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:59 vm05 bash[20951]: cluster 2026-03-10T12:33:58.039965+0000 mgr.a (mgr.14150) 68 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:59.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:33:59 vm05 bash[20951]: cluster 2026-03-10T12:33:58.039965+0000 mgr.a (mgr.14150) 68 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:59.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:59 vm01 bash[20716]: cluster 2026-03-10T12:33:58.039965+0000 mgr.a (mgr.14150) 68 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:33:59.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:33:59 vm01 bash[20716]: cluster 2026-03-10T12:33:58.039965+0000 mgr.a (mgr.14150) 68 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:01.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:01 vm05 bash[20951]: cluster 2026-03-10T12:34:00.040150+0000 mgr.a (mgr.14150) 69 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:01.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:01 vm05 bash[20951]: cluster 2026-03-10T12:34:00.040150+0000 mgr.a (mgr.14150) 69 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:01.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:01 vm01 bash[20716]: cluster 2026-03-10T12:34:00.040150+0000 mgr.a (mgr.14150) 69 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:01.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:01 vm01 bash[20716]: cluster 2026-03-10T12:34:00.040150+0000 mgr.a (mgr.14150) 69 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: cluster 2026-03-10T12:34:02.040315+0000 mgr.a (mgr.14150) 70 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: cluster 2026-03-10T12:34:02.040315+0000 mgr.a (mgr.14150) 70 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.258828+0000 mon.a (mon.0) 277 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]: dispatch 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.258828+0000 mon.a (mon.0) 277 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]: dispatch 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.258928+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.101:0/2433446223' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]: dispatch 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.258928+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.101:0/2433446223' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]: dispatch 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.261110+0000 mon.a (mon.0) 278 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]': finished 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.261110+0000 mon.a (mon.0) 278 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]': finished 2026-03-10T12:34:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: cluster 2026-03-10T12:34:02.263536+0000 mon.a (mon.0) 279 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T12:34:03.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: cluster 2026-03-10T12:34:02.263536+0000 mon.a (mon.0) 279 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T12:34:03.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.263687+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:03.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.263687+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:03.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.839646+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.101:0/148979879' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:34:03.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:03 vm05 bash[20951]: audit 2026-03-10T12:34:02.839646+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.101:0/148979879' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: cluster 2026-03-10T12:34:02.040315+0000 mgr.a (mgr.14150) 70 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: cluster 2026-03-10T12:34:02.040315+0000 mgr.a (mgr.14150) 70 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.258828+0000 mon.a (mon.0) 277 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]: dispatch 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.258828+0000 mon.a (mon.0) 277 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]: dispatch 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.258928+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.101:0/2433446223' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]: dispatch 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.258928+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.101:0/2433446223' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]: dispatch 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.261110+0000 mon.a (mon.0) 278 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]': finished 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.261110+0000 mon.a (mon.0) 278 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "dde3a957-3d01-45e3-a18f-a783cff4ef87"}]': finished 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: cluster 2026-03-10T12:34:02.263536+0000 mon.a (mon.0) 279 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: cluster 2026-03-10T12:34:02.263536+0000 mon.a (mon.0) 279 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.263687+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.263687+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.839646+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.101:0/148979879' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:34:03.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:03 vm01 bash[20716]: audit 2026-03-10T12:34:02.839646+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.101:0/148979879' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:34:05.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:05 vm05 bash[20951]: cluster 2026-03-10T12:34:04.040503+0000 mgr.a (mgr.14150) 71 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:05 vm05 bash[20951]: cluster 2026-03-10T12:34:04.040503+0000 mgr.a (mgr.14150) 71 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:05.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:05 vm01 bash[20716]: cluster 2026-03-10T12:34:04.040503+0000 mgr.a (mgr.14150) 71 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:05.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:05 vm01 bash[20716]: cluster 2026-03-10T12:34:04.040503+0000 mgr.a (mgr.14150) 71 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:07.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:07 vm05 bash[20951]: cluster 2026-03-10T12:34:06.040696+0000 mgr.a (mgr.14150) 72 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:07.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:07 vm05 bash[20951]: cluster 2026-03-10T12:34:06.040696+0000 mgr.a (mgr.14150) 72 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:07.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:07 vm01 bash[20716]: cluster 2026-03-10T12:34:06.040696+0000 mgr.a (mgr.14150) 72 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:07.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:07 vm01 bash[20716]: cluster 2026-03-10T12:34:06.040696+0000 mgr.a (mgr.14150) 72 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:09.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:08 vm05 bash[20951]: audit 2026-03-10T12:34:07.732964+0000 mon.a (mon.0) 281 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:08 vm05 bash[20951]: audit 2026-03-10T12:34:07.732964+0000 mon.a (mon.0) 281 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:08 vm05 bash[20951]: audit 2026-03-10T12:34:07.736699+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:08 vm05 bash[20951]: audit 2026-03-10T12:34:07.736699+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:08 vm05 bash[20951]: audit 2026-03-10T12:34:07.741256+0000 mon.a (mon.0) 283 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:08 vm05 bash[20951]: audit 2026-03-10T12:34:07.741256+0000 mon.a (mon.0) 283 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:08 vm01 bash[20716]: audit 2026-03-10T12:34:07.732964+0000 mon.a (mon.0) 281 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:08 vm01 bash[20716]: audit 2026-03-10T12:34:07.732964+0000 mon.a (mon.0) 281 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:08 vm01 bash[20716]: audit 2026-03-10T12:34:07.736699+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:08 vm01 bash[20716]: audit 2026-03-10T12:34:07.736699+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:08 vm01 bash[20716]: audit 2026-03-10T12:34:07.741256+0000 mon.a (mon.0) 283 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:09.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:08 vm01 bash[20716]: audit 2026-03-10T12:34:07.741256+0000 mon.a (mon.0) 283 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:10.019 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:09 vm05 bash[20951]: cluster 2026-03-10T12:34:08.040859+0000 mgr.a (mgr.14150) 73 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:10.019 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:09 vm05 bash[20951]: cluster 2026-03-10T12:34:08.040859+0000 mgr.a (mgr.14150) 73 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:10.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:09 vm01 bash[20716]: cluster 2026-03-10T12:34:08.040859+0000 mgr.a (mgr.14150) 73 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:10.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:09 vm01 bash[20716]: cluster 2026-03-10T12:34:08.040859+0000 mgr.a (mgr.14150) 73 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:11.798 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:11 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:34:11.798 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:11 vm01 bash[20716]: cluster 2026-03-10T12:34:10.041031+0000 mgr.a (mgr.14150) 74 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:11.798 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:11 vm01 bash[20716]: cluster 2026-03-10T12:34:10.041031+0000 mgr.a (mgr.14150) 74 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:11.798 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:11 vm01 bash[20716]: audit 2026-03-10T12:34:11.018925+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:34:11.798 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:11 vm01 bash[20716]: audit 2026-03-10T12:34:11.018925+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:34:11.798 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:11 vm01 bash[20716]: audit 2026-03-10T12:34:11.019346+0000 mon.a (mon.0) 285 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:11.798 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:11 vm01 bash[20716]: audit 2026-03-10T12:34:11.019346+0000 mon.a (mon.0) 285 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:11.798 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:34:11 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:34:12.071 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:34:11 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:34:12.071 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:11 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:34:12.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:11 vm05 bash[20951]: cluster 2026-03-10T12:34:10.041031+0000 mgr.a (mgr.14150) 74 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:12.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:11 vm05 bash[20951]: cluster 2026-03-10T12:34:10.041031+0000 mgr.a (mgr.14150) 74 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:12.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:11 vm05 bash[20951]: audit 2026-03-10T12:34:11.018925+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:34:12.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:11 vm05 bash[20951]: audit 2026-03-10T12:34:11.018925+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:34:12.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:11 vm05 bash[20951]: audit 2026-03-10T12:34:11.019346+0000 mon.a (mon.0) 285 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:12.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:11 vm05 bash[20951]: audit 2026-03-10T12:34:11.019346+0000 mon.a (mon.0) 285 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:13.102 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: cephadm 2026-03-10T12:34:11.019674+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Deploying daemon osd.0 on vm01 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: cephadm 2026-03-10T12:34:11.019674+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Deploying daemon osd.0 on vm01 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.034848+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.034848+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.036652+0000 mon.a (mon.0) 287 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.036652+0000 mon.a (mon.0) 287 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.037088+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.037088+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.039930+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.039930+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.043282+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.043282+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.046796+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.046796+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.520058+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.520058+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.524050+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.524050+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.528376+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.103 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:12 vm01 bash[20716]: audit 2026-03-10T12:34:12.528376+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: cephadm 2026-03-10T12:34:11.019674+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Deploying daemon osd.0 on vm01 2026-03-10T12:34:13.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: cephadm 2026-03-10T12:34:11.019674+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Deploying daemon osd.0 on vm01 2026-03-10T12:34:13.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.034848+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:13.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.034848+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:13.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.036652+0000 mon.a (mon.0) 287 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:13.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.036652+0000 mon.a (mon.0) 287 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:13.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.037088+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.037088+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.039930+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.039930+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.043282+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.043282+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.046796+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.046796+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.520058+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.520058+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.524050+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.524050+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.528376+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:13.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:12 vm05 bash[20951]: audit 2026-03-10T12:34:12.528376+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:14.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:13 vm05 bash[20951]: cluster 2026-03-10T12:34:12.041234+0000 mgr.a (mgr.14150) 76 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:14.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:13 vm05 bash[20951]: cluster 2026-03-10T12:34:12.041234+0000 mgr.a (mgr.14150) 76 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:14.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:13 vm01 bash[20716]: cluster 2026-03-10T12:34:12.041234+0000 mgr.a (mgr.14150) 76 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:14.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:13 vm01 bash[20716]: cluster 2026-03-10T12:34:12.041234+0000 mgr.a (mgr.14150) 76 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:15.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:15 vm01 bash[20716]: cluster 2026-03-10T12:34:14.041397+0000 mgr.a (mgr.14150) 77 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:15.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:15 vm01 bash[20716]: cluster 2026-03-10T12:34:14.041397+0000 mgr.a (mgr.14150) 77 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:15.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:15 vm01 bash[20716]: audit 2026-03-10T12:34:15.523255+0000 mon.a (mon.0) 295 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:34:15.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:15 vm01 bash[20716]: audit 2026-03-10T12:34:15.523255+0000 mon.a (mon.0) 295 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:34:15.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:15 vm01 bash[20716]: audit 2026-03-10T12:34:15.523504+0000 mon.b (mon.1) 8 : audit [INF] from='osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:34:15.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:15 vm01 bash[20716]: audit 2026-03-10T12:34:15.523504+0000 mon.b (mon.1) 8 : audit [INF] from='osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:34:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:15 vm05 bash[20951]: cluster 2026-03-10T12:34:14.041397+0000 mgr.a (mgr.14150) 77 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:15 vm05 bash[20951]: cluster 2026-03-10T12:34:14.041397+0000 mgr.a (mgr.14150) 77 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:15 vm05 bash[20951]: audit 2026-03-10T12:34:15.523255+0000 mon.a (mon.0) 295 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:34:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:15 vm05 bash[20951]: audit 2026-03-10T12:34:15.523255+0000 mon.a (mon.0) 295 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:34:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:15 vm05 bash[20951]: audit 2026-03-10T12:34:15.523504+0000 mon.b (mon.1) 8 : audit [INF] from='osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:34:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:15 vm05 bash[20951]: audit 2026-03-10T12:34:15.523504+0000 mon.b (mon.1) 8 : audit [INF] from='osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:15.761875+0000 mon.a (mon.0) 296 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:15.761875+0000 mon.a (mon.0) 296 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: cluster 2026-03-10T12:34:15.764682+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: cluster 2026-03-10T12:34:15.764682+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:15.765403+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:15.765403+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:15.765508+0000 mon.a (mon.0) 299 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:15.765508+0000 mon.a (mon.0) 299 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:15.765842+0000 mon.b (mon.1) 9 : audit [INF] from='osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:15.765842+0000 mon.b (mon.1) 9 : audit [INF] from='osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:16.765170+0000 mon.a (mon.0) 300 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: audit 2026-03-10T12:34:16.765170+0000 mon.a (mon.0) 300 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: cluster 2026-03-10T12:34:16.767978+0000 mon.a (mon.0) 301 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T12:34:17.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:16 vm05 bash[20951]: cluster 2026-03-10T12:34:16.767978+0000 mon.a (mon.0) 301 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T12:34:17.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:15.761875+0000 mon.a (mon.0) 296 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T12:34:17.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:15.761875+0000 mon.a (mon.0) 296 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T12:34:17.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: cluster 2026-03-10T12:34:15.764682+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T12:34:17.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: cluster 2026-03-10T12:34:15.764682+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T12:34:17.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:15.765403+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:17.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:15.765403+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:17.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:15.765508+0000 mon.a (mon.0) 299 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T12:34:17.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:15.765508+0000 mon.a (mon.0) 299 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T12:34:17.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:15.765842+0000 mon.b (mon.1) 9 : audit [INF] from='osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T12:34:17.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:15.765842+0000 mon.b (mon.1) 9 : audit [INF] from='osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T12:34:17.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:16.765170+0000 mon.a (mon.0) 300 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T12:34:17.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: audit 2026-03-10T12:34:16.765170+0000 mon.a (mon.0) 300 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T12:34:17.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: cluster 2026-03-10T12:34:16.767978+0000 mon.a (mon.0) 301 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T12:34:17.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:16 vm01 bash[20716]: cluster 2026-03-10T12:34:16.767978+0000 mon.a (mon.0) 301 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T12:34:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:17 vm05 bash[20951]: cluster 2026-03-10T12:34:16.041597+0000 mgr.a (mgr.14150) 78 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:17 vm05 bash[20951]: cluster 2026-03-10T12:34:16.041597+0000 mgr.a (mgr.14150) 78 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:17 vm05 bash[20951]: audit 2026-03-10T12:34:16.768378+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:17 vm05 bash[20951]: audit 2026-03-10T12:34:16.768378+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:17 vm05 bash[20951]: audit 2026-03-10T12:34:16.774691+0000 mon.a (mon.0) 303 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:17 vm05 bash[20951]: audit 2026-03-10T12:34:16.774691+0000 mon.a (mon.0) 303 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:17 vm05 bash[20951]: audit 2026-03-10T12:34:17.770950+0000 mon.a (mon.0) 304 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:17 vm05 bash[20951]: audit 2026-03-10T12:34:17.770950+0000 mon.a (mon.0) 304 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:17 vm01 bash[20716]: cluster 2026-03-10T12:34:16.041597+0000 mgr.a (mgr.14150) 78 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:17 vm01 bash[20716]: cluster 2026-03-10T12:34:16.041597+0000 mgr.a (mgr.14150) 78 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:17 vm01 bash[20716]: audit 2026-03-10T12:34:16.768378+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:17 vm01 bash[20716]: audit 2026-03-10T12:34:16.768378+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:17 vm01 bash[20716]: audit 2026-03-10T12:34:16.774691+0000 mon.a (mon.0) 303 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:17 vm01 bash[20716]: audit 2026-03-10T12:34:16.774691+0000 mon.a (mon.0) 303 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:17 vm01 bash[20716]: audit 2026-03-10T12:34:17.770950+0000 mon.a (mon.0) 304 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:18.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:17 vm01 bash[20716]: audit 2026-03-10T12:34:17.770950+0000 mon.a (mon.0) 304 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:19.172 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 0 on host 'vm01' 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: cluster 2026-03-10T12:34:16.475516+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: cluster 2026-03-10T12:34:16.475516+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: cluster 2026-03-10T12:34:16.475583+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: cluster 2026-03-10T12:34:16.475583+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: cluster 2026-03-10T12:34:17.780835+0000 mon.a (mon.0) 305 : cluster [INF] osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180] boot 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: cluster 2026-03-10T12:34:17.780835+0000 mon.a (mon.0) 305 : cluster [INF] osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180] boot 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: cluster 2026-03-10T12:34:17.780928+0000 mon.a (mon.0) 306 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: cluster 2026-03-10T12:34:17.780928+0000 mon.a (mon.0) 306 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: audit 2026-03-10T12:34:17.781477+0000 mon.a (mon.0) 307 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:19.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:18 vm05 bash[20951]: audit 2026-03-10T12:34:17.781477+0000 mon.a (mon.0) 307 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: cluster 2026-03-10T12:34:16.475516+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: cluster 2026-03-10T12:34:16.475516+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: cluster 2026-03-10T12:34:16.475583+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: cluster 2026-03-10T12:34:16.475583+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: cluster 2026-03-10T12:34:17.780835+0000 mon.a (mon.0) 305 : cluster [INF] osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180] boot 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: cluster 2026-03-10T12:34:17.780835+0000 mon.a (mon.0) 305 : cluster [INF] osd.0 [v2:192.168.123.101:6802/2723523180,v1:192.168.123.101:6803/2723523180] boot 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: cluster 2026-03-10T12:34:17.780928+0000 mon.a (mon.0) 306 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: cluster 2026-03-10T12:34:17.780928+0000 mon.a (mon.0) 306 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: audit 2026-03-10T12:34:17.781477+0000 mon.a (mon.0) 307 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:19.182 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:18 vm01 bash[20716]: audit 2026-03-10T12:34:17.781477+0000 mon.a (mon.0) 307 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:34:19.239 DEBUG:teuthology.orchestra.run.vm01:osd.0> sudo journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@osd.0.service 2026-03-10T12:34:19.240 INFO:tasks.cephadm:Deploying osd.1 on vm05 with /dev/vde... 2026-03-10T12:34:19.240 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- lvm zap /dev/vde 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: cluster 2026-03-10T12:34:18.041776+0000 mgr.a (mgr.14150) 79 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: cluster 2026-03-10T12:34:18.041776+0000 mgr.a (mgr.14150) 79 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: cluster 2026-03-10T12:34:18.794542+0000 mon.a (mon.0) 308 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: cluster 2026-03-10T12:34:18.794542+0000 mon.a (mon.0) 308 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:18.882961+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:18.882961+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:18.944728+0000 mon.a (mon.0) 310 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:18.944728+0000 mon.a (mon.0) 310 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.065014+0000 mon.a (mon.0) 311 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.065014+0000 mon.a (mon.0) 311 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.113572+0000 mon.a (mon.0) 312 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.113572+0000 mon.a (mon.0) 312 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.114519+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.114519+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.169887+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.169887+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.176196+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:19 vm05 bash[20951]: audit 2026-03-10T12:34:19.176196+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: cluster 2026-03-10T12:34:18.041776+0000 mgr.a (mgr.14150) 79 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:20.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: cluster 2026-03-10T12:34:18.041776+0000 mgr.a (mgr.14150) 79 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:34:20.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: cluster 2026-03-10T12:34:18.794542+0000 mon.a (mon.0) 308 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T12:34:20.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: cluster 2026-03-10T12:34:18.794542+0000 mon.a (mon.0) 308 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T12:34:20.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:18.882961+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:18.882961+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:18.944728+0000 mon.a (mon.0) 310 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:18.944728+0000 mon.a (mon.0) 310 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.065014+0000 mon.a (mon.0) 311 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.065014+0000 mon.a (mon.0) 311 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.113572+0000 mon.a (mon.0) 312 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.113572+0000 mon.a (mon.0) 312 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.114519+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.114519+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.169887+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.169887+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.176196+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:20.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:19 vm01 bash[20716]: audit 2026-03-10T12:34:19.176196+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:22.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:21 vm05 bash[20951]: cluster 2026-03-10T12:34:20.042009+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:22.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:21 vm05 bash[20951]: cluster 2026-03-10T12:34:20.042009+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:22.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:21 vm01 bash[20716]: cluster 2026-03-10T12:34:20.042009+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:22.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:21 vm01 bash[20716]: cluster 2026-03-10T12:34:20.042009+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:23.854 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.b/config 2026-03-10T12:34:24.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:23 vm05 bash[20951]: cluster 2026-03-10T12:34:22.042223+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:24.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:23 vm05 bash[20951]: cluster 2026-03-10T12:34:22.042223+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:24.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:23 vm01 bash[20716]: cluster 2026-03-10T12:34:22.042223+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:24.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:23 vm01 bash[20716]: cluster 2026-03-10T12:34:22.042223+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:24.687 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T12:34:24.702 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch daemon add osd vm05:/dev/vde 2026-03-10T12:34:26.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:25 vm05 bash[20951]: cluster 2026-03-10T12:34:24.042448+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:26.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:25 vm05 bash[20951]: cluster 2026-03-10T12:34:24.042448+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:26.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:25 vm01 bash[20716]: cluster 2026-03-10T12:34:24.042448+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:26.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:25 vm01 bash[20716]: cluster 2026-03-10T12:34:24.042448+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:27.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:27 vm05 bash[20951]: cluster 2026-03-10T12:34:26.042644+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:27.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:27 vm05 bash[20951]: cluster 2026-03-10T12:34:26.042644+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:27.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:27 vm01 bash[20716]: cluster 2026-03-10T12:34:26.042644+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:27.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:27 vm01 bash[20716]: cluster 2026-03-10T12:34:26.042644+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:29.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.812723+0000 mon.a (mon.0) 316 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.812723+0000 mon.a (mon.0) 316 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.817044+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.817044+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: cephadm 2026-03-10T12:34:27.818035+0000 mgr.a (mgr.14150) 84 : cephadm [INF] Detected new or changed devices on vm01 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: cephadm 2026-03-10T12:34:27.818035+0000 mgr.a (mgr.14150) 84 : cephadm [INF] Detected new or changed devices on vm01 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.818563+0000 mon.a (mon.0) 318 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.818563+0000 mon.a (mon.0) 318 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.820301+0000 mon.a (mon.0) 319 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.820301+0000 mon.a (mon.0) 319 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.820782+0000 mon.a (mon.0) 320 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.820782+0000 mon.a (mon.0) 320 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.822355+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.822355+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.827575+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:28 vm05 bash[20951]: audit 2026-03-10T12:34:27.827575+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.812723+0000 mon.a (mon.0) 316 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.812723+0000 mon.a (mon.0) 316 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.817044+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.817044+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: cephadm 2026-03-10T12:34:27.818035+0000 mgr.a (mgr.14150) 84 : cephadm [INF] Detected new or changed devices on vm01 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: cephadm 2026-03-10T12:34:27.818035+0000 mgr.a (mgr.14150) 84 : cephadm [INF] Detected new or changed devices on vm01 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.818563+0000 mon.a (mon.0) 318 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.818563+0000 mon.a (mon.0) 318 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:29.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.820301+0000 mon.a (mon.0) 319 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:29.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.820301+0000 mon.a (mon.0) 319 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:29.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.820782+0000 mon.a (mon.0) 320 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:29.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.820782+0000 mon.a (mon.0) 320 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:29.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.822355+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.822355+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.827575+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:28 vm01 bash[20716]: audit 2026-03-10T12:34:27.827575+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:29.312 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.b/config 2026-03-10T12:34:30.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:29 vm05 bash[20951]: cluster 2026-03-10T12:34:28.042860+0000 mgr.a (mgr.14150) 85 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:30.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:29 vm05 bash[20951]: cluster 2026-03-10T12:34:28.042860+0000 mgr.a (mgr.14150) 85 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:29 vm05 bash[20951]: audit 2026-03-10T12:34:29.827030+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:34:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:29 vm05 bash[20951]: audit 2026-03-10T12:34:29.827030+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:34:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:29 vm05 bash[20951]: audit 2026-03-10T12:34:29.828629+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:34:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:29 vm05 bash[20951]: audit 2026-03-10T12:34:29.828629+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:34:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:29 vm05 bash[20951]: audit 2026-03-10T12:34:29.829364+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:30.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:29 vm05 bash[20951]: audit 2026-03-10T12:34:29.829364+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:30.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:29 vm01 bash[20716]: cluster 2026-03-10T12:34:28.042860+0000 mgr.a (mgr.14150) 85 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:30.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:29 vm01 bash[20716]: cluster 2026-03-10T12:34:28.042860+0000 mgr.a (mgr.14150) 85 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:30.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:29 vm01 bash[20716]: audit 2026-03-10T12:34:29.827030+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:34:30.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:29 vm01 bash[20716]: audit 2026-03-10T12:34:29.827030+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:34:30.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:29 vm01 bash[20716]: audit 2026-03-10T12:34:29.828629+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:34:30.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:29 vm01 bash[20716]: audit 2026-03-10T12:34:29.828629+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:34:30.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:29 vm01 bash[20716]: audit 2026-03-10T12:34:29.829364+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:30.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:29 vm01 bash[20716]: audit 2026-03-10T12:34:29.829364+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:31.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:30 vm05 bash[20951]: audit 2026-03-10T12:34:29.825426+0000 mgr.a (mgr.14150) 86 : audit [DBG] from='client.24123 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:34:31.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:30 vm05 bash[20951]: audit 2026-03-10T12:34:29.825426+0000 mgr.a (mgr.14150) 86 : audit [DBG] from='client.24123 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:34:31.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:30 vm01 bash[20716]: audit 2026-03-10T12:34:29.825426+0000 mgr.a (mgr.14150) 86 : audit [DBG] from='client.24123 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:34:31.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:30 vm01 bash[20716]: audit 2026-03-10T12:34:29.825426+0000 mgr.a (mgr.14150) 86 : audit [DBG] from='client.24123 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: cluster 2026-03-10T12:34:30.043060+0000 mgr.a (mgr.14150) 87 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: cluster 2026-03-10T12:34:30.043060+0000 mgr.a (mgr.14150) 87 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.372924+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.372924+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.376859+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.376859+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.378226+0000 mon.a (mon.0) 328 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.378226+0000 mon.a (mon.0) 328 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.380384+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.380384+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.382148+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.382148+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.382471+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.382471+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.387695+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:31 vm05 bash[20951]: audit 2026-03-10T12:34:31.387695+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: cluster 2026-03-10T12:34:30.043060+0000 mgr.a (mgr.14150) 87 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: cluster 2026-03-10T12:34:30.043060+0000 mgr.a (mgr.14150) 87 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.372924+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.372924+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.376859+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.376859+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.378226+0000 mon.a (mon.0) 328 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.378226+0000 mon.a (mon.0) 328 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:32.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.380384+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:32.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.380384+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:32.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.382148+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.382148+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.382471+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:32.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.382471+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:32.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.387695+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:32.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:31 vm01 bash[20716]: audit 2026-03-10T12:34:31.387695+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:33 vm05 bash[20951]: cluster 2026-03-10T12:34:32.043313+0000 mgr.a (mgr.14150) 88 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:33.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:33 vm05 bash[20951]: cluster 2026-03-10T12:34:32.043313+0000 mgr.a (mgr.14150) 88 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:33.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:33 vm05 bash[20951]: audit 2026-03-10T12:34:32.587992+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:33 vm05 bash[20951]: audit 2026-03-10T12:34:32.587992+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:33 vm05 bash[20951]: audit 2026-03-10T12:34:32.592342+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:33 vm05 bash[20951]: audit 2026-03-10T12:34:32.592342+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:33 vm05 bash[20951]: audit 2026-03-10T12:34:32.597071+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:33 vm05 bash[20951]: audit 2026-03-10T12:34:32.597071+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:33 vm01 bash[20716]: cluster 2026-03-10T12:34:32.043313+0000 mgr.a (mgr.14150) 88 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:33.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:33 vm01 bash[20716]: cluster 2026-03-10T12:34:32.043313+0000 mgr.a (mgr.14150) 88 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:33.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:33 vm01 bash[20716]: audit 2026-03-10T12:34:32.587992+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:33 vm01 bash[20716]: audit 2026-03-10T12:34:32.587992+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:33 vm01 bash[20716]: audit 2026-03-10T12:34:32.592342+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:33 vm01 bash[20716]: audit 2026-03-10T12:34:32.592342+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:33 vm01 bash[20716]: audit 2026-03-10T12:34:32.597071+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:33.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:33 vm01 bash[20716]: audit 2026-03-10T12:34:32.597071+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:35.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: cluster 2026-03-10T12:34:34.043556+0000 mgr.a (mgr.14150) 89 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:35.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: cluster 2026-03-10T12:34:34.043556+0000 mgr.a (mgr.14150) 89 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: audit 2026-03-10T12:34:35.187791+0000 mon.a (mon.0) 336 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]: dispatch 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: audit 2026-03-10T12:34:35.187791+0000 mon.a (mon.0) 336 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]: dispatch 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: audit 2026-03-10T12:34:35.187939+0000 mon.b (mon.1) 10 : audit [INF] from='client.? 192.168.123.105:0/2807268425' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]: dispatch 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: audit 2026-03-10T12:34:35.187939+0000 mon.b (mon.1) 10 : audit [INF] from='client.? 192.168.123.105:0/2807268425' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]: dispatch 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: audit 2026-03-10T12:34:35.190345+0000 mon.a (mon.0) 337 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]': finished 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: audit 2026-03-10T12:34:35.190345+0000 mon.a (mon.0) 337 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]': finished 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: cluster 2026-03-10T12:34:35.192905+0000 mon.a (mon.0) 338 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: cluster 2026-03-10T12:34:35.192905+0000 mon.a (mon.0) 338 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: audit 2026-03-10T12:34:35.193024+0000 mon.a (mon.0) 339 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:35.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:35 vm05 bash[20951]: audit 2026-03-10T12:34:35.193024+0000 mon.a (mon.0) 339 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: cluster 2026-03-10T12:34:34.043556+0000 mgr.a (mgr.14150) 89 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: cluster 2026-03-10T12:34:34.043556+0000 mgr.a (mgr.14150) 89 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: audit 2026-03-10T12:34:35.187791+0000 mon.a (mon.0) 336 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]: dispatch 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: audit 2026-03-10T12:34:35.187791+0000 mon.a (mon.0) 336 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]: dispatch 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: audit 2026-03-10T12:34:35.187939+0000 mon.b (mon.1) 10 : audit [INF] from='client.? 192.168.123.105:0/2807268425' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]: dispatch 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: audit 2026-03-10T12:34:35.187939+0000 mon.b (mon.1) 10 : audit [INF] from='client.? 192.168.123.105:0/2807268425' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]: dispatch 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: audit 2026-03-10T12:34:35.190345+0000 mon.a (mon.0) 337 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]': finished 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: audit 2026-03-10T12:34:35.190345+0000 mon.a (mon.0) 337 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "45acbc28-d814-49a2-a229-15ef9cf0619b"}]': finished 2026-03-10T12:34:35.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: cluster 2026-03-10T12:34:35.192905+0000 mon.a (mon.0) 338 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T12:34:35.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: cluster 2026-03-10T12:34:35.192905+0000 mon.a (mon.0) 338 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T12:34:35.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: audit 2026-03-10T12:34:35.193024+0000 mon.a (mon.0) 339 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:35.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:35 vm01 bash[20716]: audit 2026-03-10T12:34:35.193024+0000 mon.a (mon.0) 339 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:36.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:36 vm05 bash[20951]: audit 2026-03-10T12:34:35.795650+0000 mon.b (mon.1) 11 : audit [DBG] from='client.? 192.168.123.105:0/1971454927' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:34:36.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:36 vm05 bash[20951]: audit 2026-03-10T12:34:35.795650+0000 mon.b (mon.1) 11 : audit [DBG] from='client.? 192.168.123.105:0/1971454927' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:34:36.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:36 vm01 bash[20716]: audit 2026-03-10T12:34:35.795650+0000 mon.b (mon.1) 11 : audit [DBG] from='client.? 192.168.123.105:0/1971454927' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:34:36.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:36 vm01 bash[20716]: audit 2026-03-10T12:34:35.795650+0000 mon.b (mon.1) 11 : audit [DBG] from='client.? 192.168.123.105:0/1971454927' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:34:37.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:37 vm05 bash[20951]: cluster 2026-03-10T12:34:36.043750+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:37.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:37 vm05 bash[20951]: cluster 2026-03-10T12:34:36.043750+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:37.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:37 vm01 bash[20716]: cluster 2026-03-10T12:34:36.043750+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:37.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:37 vm01 bash[20716]: cluster 2026-03-10T12:34:36.043750+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:39.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:39 vm05 bash[20951]: cluster 2026-03-10T12:34:38.043958+0000 mgr.a (mgr.14150) 91 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:39.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:39 vm05 bash[20951]: cluster 2026-03-10T12:34:38.043958+0000 mgr.a (mgr.14150) 91 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:39.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:39 vm01 bash[20716]: cluster 2026-03-10T12:34:38.043958+0000 mgr.a (mgr.14150) 91 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:39.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:39 vm01 bash[20716]: cluster 2026-03-10T12:34:38.043958+0000 mgr.a (mgr.14150) 91 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:41.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:41 vm05 bash[20951]: cluster 2026-03-10T12:34:40.044202+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:41.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:41 vm05 bash[20951]: cluster 2026-03-10T12:34:40.044202+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:41.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:41 vm01 bash[20716]: cluster 2026-03-10T12:34:40.044202+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:41.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:41 vm01 bash[20716]: cluster 2026-03-10T12:34:40.044202+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:43.873 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:43 vm05 bash[20951]: cluster 2026-03-10T12:34:42.044490+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:43.873 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:43 vm05 bash[20951]: cluster 2026-03-10T12:34:42.044490+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:43.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:43 vm01 bash[20716]: cluster 2026-03-10T12:34:42.044490+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:43.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:43 vm01 bash[20716]: cluster 2026-03-10T12:34:42.044490+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:44.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:44 vm05 bash[20951]: audit 2026-03-10T12:34:44.231619+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:34:44.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:44 vm05 bash[20951]: audit 2026-03-10T12:34:44.231619+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:34:44.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:44 vm05 bash[20951]: audit 2026-03-10T12:34:44.232227+0000 mon.a (mon.0) 341 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:44.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:44 vm05 bash[20951]: audit 2026-03-10T12:34:44.232227+0000 mon.a (mon.0) 341 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:44.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:44 vm01 bash[20716]: audit 2026-03-10T12:34:44.231619+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:34:44.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:44 vm01 bash[20716]: audit 2026-03-10T12:34:44.231619+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:34:44.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:44 vm01 bash[20716]: audit 2026-03-10T12:34:44.232227+0000 mon.a (mon.0) 341 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:44.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:44 vm01 bash[20716]: audit 2026-03-10T12:34:44.232227+0000 mon.a (mon.0) 341 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:45.048 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:44 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:34:45.049 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:34:44 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:34:45.335 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:34:45 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:34:45.335 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: cluster 2026-03-10T12:34:44.044695+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: cluster 2026-03-10T12:34:44.044695+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: cephadm 2026-03-10T12:34:44.232701+0000 mgr.a (mgr.14150) 95 : cephadm [INF] Deploying daemon osd.1 on vm05 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: cephadm 2026-03-10T12:34:44.232701+0000 mgr.a (mgr.14150) 95 : cephadm [INF] Deploying daemon osd.1 on vm05 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.313588+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.313588+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.315917+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.315917+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.316552+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.316552+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:45.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.318573+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.318573+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.323063+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.323063+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.330957+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:45 vm05 bash[20951]: audit 2026-03-10T12:34:45.330957+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: cluster 2026-03-10T12:34:44.044695+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:45.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: cluster 2026-03-10T12:34:44.044695+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:45.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: cephadm 2026-03-10T12:34:44.232701+0000 mgr.a (mgr.14150) 95 : cephadm [INF] Deploying daemon osd.1 on vm05 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: cephadm 2026-03-10T12:34:44.232701+0000 mgr.a (mgr.14150) 95 : cephadm [INF] Deploying daemon osd.1 on vm05 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.313588+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.313588+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.315917+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.315917+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.316552+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.316552+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.318573+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.318573+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.323063+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.323063+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.330957+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:45.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:45 vm01 bash[20716]: audit 2026-03-10T12:34:45.330957+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:47.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:47 vm05 bash[20951]: cluster 2026-03-10T12:34:46.044984+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:47.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:47 vm05 bash[20951]: cluster 2026-03-10T12:34:46.044984+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:47.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:47 vm01 bash[20716]: cluster 2026-03-10T12:34:46.044984+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:47.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:47 vm01 bash[20716]: cluster 2026-03-10T12:34:46.044984+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:49.142 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:48 vm05 bash[20951]: audit 2026-03-10T12:34:48.485495+0000 mon.a (mon.0) 348 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T12:34:49.142 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:48 vm05 bash[20951]: audit 2026-03-10T12:34:48.485495+0000 mon.a (mon.0) 348 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T12:34:49.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:48 vm01 bash[20716]: audit 2026-03-10T12:34:48.485495+0000 mon.a (mon.0) 348 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T12:34:49.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:48 vm01 bash[20716]: audit 2026-03-10T12:34:48.485495+0000 mon.a (mon.0) 348 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T12:34:50.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: cluster 2026-03-10T12:34:48.045202+0000 mgr.a (mgr.14150) 97 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: cluster 2026-03-10T12:34:48.045202+0000 mgr.a (mgr.14150) 97 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: audit 2026-03-10T12:34:48.932976+0000 mon.a (mon.0) 349 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: audit 2026-03-10T12:34:48.932976+0000 mon.a (mon.0) 349 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: cluster 2026-03-10T12:34:48.965469+0000 mon.a (mon.0) 350 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: cluster 2026-03-10T12:34:48.965469+0000 mon.a (mon.0) 350 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: audit 2026-03-10T12:34:48.965685+0000 mon.a (mon.0) 351 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: audit 2026-03-10T12:34:48.965685+0000 mon.a (mon.0) 351 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: audit 2026-03-10T12:34:48.965794+0000 mon.a (mon.0) 352 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:50.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:49 vm05 bash[20951]: audit 2026-03-10T12:34:48.965794+0000 mon.a (mon.0) 352 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:50.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: cluster 2026-03-10T12:34:48.045202+0000 mgr.a (mgr.14150) 97 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:50.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: cluster 2026-03-10T12:34:48.045202+0000 mgr.a (mgr.14150) 97 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:50.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: audit 2026-03-10T12:34:48.932976+0000 mon.a (mon.0) 349 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T12:34:50.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: audit 2026-03-10T12:34:48.932976+0000 mon.a (mon.0) 349 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T12:34:50.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: cluster 2026-03-10T12:34:48.965469+0000 mon.a (mon.0) 350 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T12:34:50.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: cluster 2026-03-10T12:34:48.965469+0000 mon.a (mon.0) 350 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T12:34:50.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: audit 2026-03-10T12:34:48.965685+0000 mon.a (mon.0) 351 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T12:34:50.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: audit 2026-03-10T12:34:48.965685+0000 mon.a (mon.0) 351 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T12:34:50.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: audit 2026-03-10T12:34:48.965794+0000 mon.a (mon.0) 352 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:50.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:49 vm01 bash[20716]: audit 2026-03-10T12:34:48.965794+0000 mon.a (mon.0) 352 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: audit 2026-03-10T12:34:49.938172+0000 mon.a (mon.0) 353 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: audit 2026-03-10T12:34:49.938172+0000 mon.a (mon.0) 353 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: cluster 2026-03-10T12:34:49.940096+0000 mon.a (mon.0) 354 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: cluster 2026-03-10T12:34:49.940096+0000 mon.a (mon.0) 354 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: audit 2026-03-10T12:34:49.940927+0000 mon.a (mon.0) 355 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: audit 2026-03-10T12:34:49.940927+0000 mon.a (mon.0) 355 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: audit 2026-03-10T12:34:49.943991+0000 mon.a (mon.0) 356 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: audit 2026-03-10T12:34:49.943991+0000 mon.a (mon.0) 356 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: cluster 2026-03-10T12:34:50.045402+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:51.689 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:51 vm05 bash[20951]: cluster 2026-03-10T12:34:50.045402+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:51.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: audit 2026-03-10T12:34:49.938172+0000 mon.a (mon.0) 353 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: audit 2026-03-10T12:34:49.938172+0000 mon.a (mon.0) 353 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: cluster 2026-03-10T12:34:49.940096+0000 mon.a (mon.0) 354 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: cluster 2026-03-10T12:34:49.940096+0000 mon.a (mon.0) 354 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: audit 2026-03-10T12:34:49.940927+0000 mon.a (mon.0) 355 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: audit 2026-03-10T12:34:49.940927+0000 mon.a (mon.0) 355 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: audit 2026-03-10T12:34:49.943991+0000 mon.a (mon.0) 356 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: audit 2026-03-10T12:34:49.943991+0000 mon.a (mon.0) 356 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: cluster 2026-03-10T12:34:50.045402+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:51.738 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:51 vm01 bash[20716]: cluster 2026-03-10T12:34:50.045402+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: cluster 2026-03-10T12:34:49.496985+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: cluster 2026-03-10T12:34:49.496985+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: cluster 2026-03-10T12:34:49.497046+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: cluster 2026-03-10T12:34:49.497046+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:50.975095+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:50.975095+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.016958+0000 mon.a (mon.0) 358 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.016958+0000 mon.a (mon.0) 358 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.220604+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.220604+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.226537+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.226537+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.299624+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.299624+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.944384+0000 mon.a (mon.0) 362 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:52.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:52 vm05 bash[20951]: audit 2026-03-10T12:34:51.944384+0000 mon.a (mon.0) 362 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:52.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: cluster 2026-03-10T12:34:49.496985+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T12:34:52.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: cluster 2026-03-10T12:34:49.496985+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T12:34:52.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: cluster 2026-03-10T12:34:49.497046+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:34:52.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: cluster 2026-03-10T12:34:49.497046+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:34:52.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:50.975095+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:50.975095+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.016958+0000 mon.a (mon.0) 358 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.016958+0000 mon.a (mon.0) 358 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.220604+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.220604+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.226537+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.226537+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.299624+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.299624+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692]' entity='osd.1' 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.944384+0000 mon.a (mon.0) 362 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:52.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:52 vm01 bash[20716]: audit 2026-03-10T12:34:51.944384+0000 mon.a (mon.0) 362 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:53.451 INFO:teuthology.orchestra.run.vm05.stdout:Created osd(s) 1 on host 'vm05' 2026-03-10T12:34:53.520 DEBUG:teuthology.orchestra.run.vm05:osd.1> sudo journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@osd.1.service 2026-03-10T12:34:53.521 INFO:tasks.cephadm:Waiting for 2 OSDs to come up... 2026-03-10T12:34:53.522 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph osd stat -f json 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: cluster 2026-03-10T12:34:52.045640+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: cluster 2026-03-10T12:34:52.045640+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: cluster 2026-03-10T12:34:52.379414+0000 mon.a (mon.0) 363 : cluster [INF] osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692] boot 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: cluster 2026-03-10T12:34:52.379414+0000 mon.a (mon.0) 363 : cluster [INF] osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692] boot 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: cluster 2026-03-10T12:34:52.379489+0000 mon.a (mon.0) 364 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: cluster 2026-03-10T12:34:52.379489+0000 mon.a (mon.0) 364 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.379796+0000 mon.a (mon.0) 365 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.379796+0000 mon.a (mon.0) 365 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.683226+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.683226+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.687368+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.687368+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: cephadm 2026-03-10T12:34:52.688240+0000 mgr.a (mgr.14150) 100 : cephadm [INF] Detected new or changed devices on vm05 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: cephadm 2026-03-10T12:34:52.688240+0000 mgr.a (mgr.14150) 100 : cephadm [INF] Detected new or changed devices on vm05 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.688618+0000 mon.a (mon.0) 368 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.688618+0000 mon.a (mon.0) 368 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.690223+0000 mon.a (mon.0) 369 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.690223+0000 mon.a (mon.0) 369 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.690700+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.690700+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.692880+0000 mon.a (mon.0) 371 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.692880+0000 mon.a (mon.0) 371 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.696297+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:53 vm05 bash[20951]: audit 2026-03-10T12:34:52.696297+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: cluster 2026-03-10T12:34:52.045640+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: cluster 2026-03-10T12:34:52.045640+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: cluster 2026-03-10T12:34:52.379414+0000 mon.a (mon.0) 363 : cluster [INF] osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692] boot 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: cluster 2026-03-10T12:34:52.379414+0000 mon.a (mon.0) 363 : cluster [INF] osd.1 [v2:192.168.123.105:6800/3430972692,v1:192.168.123.105:6801/3430972692] boot 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: cluster 2026-03-10T12:34:52.379489+0000 mon.a (mon.0) 364 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: cluster 2026-03-10T12:34:52.379489+0000 mon.a (mon.0) 364 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.379796+0000 mon.a (mon.0) 365 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.379796+0000 mon.a (mon.0) 365 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.683226+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.683226+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.687368+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.687368+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: cephadm 2026-03-10T12:34:52.688240+0000 mgr.a (mgr.14150) 100 : cephadm [INF] Detected new or changed devices on vm05 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: cephadm 2026-03-10T12:34:52.688240+0000 mgr.a (mgr.14150) 100 : cephadm [INF] Detected new or changed devices on vm05 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.688618+0000 mon.a (mon.0) 368 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.688618+0000 mon.a (mon.0) 368 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.690223+0000 mon.a (mon.0) 369 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.690223+0000 mon.a (mon.0) 369 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.690700+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.690700+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.692880+0000 mon.a (mon.0) 371 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.692880+0000 mon.a (mon.0) 371 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.696297+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:53 vm01 bash[20716]: audit 2026-03-10T12:34:52.696297+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: cluster 2026-03-10T12:34:53.395769+0000 mon.a (mon.0) 373 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: cluster 2026-03-10T12:34:53.395769+0000 mon.a (mon.0) 373 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.440691+0000 mon.a (mon.0) 374 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.440691+0000 mon.a (mon.0) 374 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.442703+0000 mon.a (mon.0) 375 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.442703+0000 mon.a (mon.0) 375 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.445194+0000 mon.a (mon.0) 376 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.445194+0000 mon.a (mon.0) 376 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.446921+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.446921+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.447738+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.447738+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.449502+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.449502+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.452943+0000 mon.a (mon.0) 380 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:53.452943+0000 mon.a (mon.0) 380 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.223858+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.223858+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.278306+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.278306+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.311126+0000 mon.a (mon.0) 383 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.311126+0000 mon.a (mon.0) 383 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.348289+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.348289+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.348663+0000 mon.a (mon.0) 385 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.348663+0000 mon.a (mon.0) 385 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.370360+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.370360+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.391240+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:54 vm05 bash[20951]: audit 2026-03-10T12:34:54.391240+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: cluster 2026-03-10T12:34:53.395769+0000 mon.a (mon.0) 373 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: cluster 2026-03-10T12:34:53.395769+0000 mon.a (mon.0) 373 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.440691+0000 mon.a (mon.0) 374 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.440691+0000 mon.a (mon.0) 374 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.442703+0000 mon.a (mon.0) 375 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.442703+0000 mon.a (mon.0) 375 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.445194+0000 mon.a (mon.0) 376 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.445194+0000 mon.a (mon.0) 376 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.446921+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.446921+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.447738+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.447738+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.449502+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.449502+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.452943+0000 mon.a (mon.0) 380 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:53.452943+0000 mon.a (mon.0) 380 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.223858+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.223858+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.278306+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.278306+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.311126+0000 mon.a (mon.0) 383 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.311126+0000 mon.a (mon.0) 383 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.348289+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.348289+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.348663+0000 mon.a (mon.0) 385 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.348663+0000 mon.a (mon.0) 385 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.370360+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.370360+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.391240+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:54.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:54 vm01 bash[20716]: audit 2026-03-10T12:34:54.391240+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:34:55.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:55 vm05 bash[20951]: cluster 2026-03-10T12:34:54.045852+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:55.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:55 vm05 bash[20951]: cluster 2026-03-10T12:34:54.045852+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:55.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:55 vm01 bash[20716]: cluster 2026-03-10T12:34:54.045852+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:55.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:55 vm01 bash[20716]: cluster 2026-03-10T12:34:54.045852+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:57.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:57 vm05 bash[20951]: cluster 2026-03-10T12:34:56.046034+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:57.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:57 vm05 bash[20951]: cluster 2026-03-10T12:34:56.046034+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:57.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:57 vm01 bash[20716]: cluster 2026-03-10T12:34:56.046034+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:57.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:57 vm01 bash[20716]: cluster 2026-03-10T12:34:56.046034+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:58.138 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:34:58.363 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:34:58.410 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"num_osds":2,"num_up_osds":2,"osd_up_since":1773146092,"num_in_osds":2,"osd_in_since":1773146075,"num_remapped_pgs":0} 2026-03-10T12:34:58.410 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph osd dump --format=json 2026-03-10T12:34:58.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:58 vm01 bash[20716]: audit 2026-03-10T12:34:58.362739+0000 mon.a (mon.0) 388 : audit [DBG] from='client.? 192.168.123.101:0/4044481233' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:34:58.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:58 vm01 bash[20716]: audit 2026-03-10T12:34:58.362739+0000 mon.a (mon.0) 388 : audit [DBG] from='client.? 192.168.123.101:0/4044481233' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:34:58.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:58 vm05 bash[20951]: audit 2026-03-10T12:34:58.362739+0000 mon.a (mon.0) 388 : audit [DBG] from='client.? 192.168.123.101:0/4044481233' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:34:58.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:58 vm05 bash[20951]: audit 2026-03-10T12:34:58.362739+0000 mon.a (mon.0) 388 : audit [DBG] from='client.? 192.168.123.101:0/4044481233' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:34:59.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:59 vm05 bash[20951]: cluster 2026-03-10T12:34:58.046189+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:59.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:34:59 vm05 bash[20951]: cluster 2026-03-10T12:34:58.046189+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:59.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:59 vm01 bash[20716]: cluster 2026-03-10T12:34:58.046189+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:34:59.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:34:59 vm01 bash[20716]: cluster 2026-03-10T12:34:58.046189+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:01.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:01 vm01 bash[20716]: cluster 2026-03-10T12:35:00.046409+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:01.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:01 vm01 bash[20716]: cluster 2026-03-10T12:35:00.046409+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:01.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:01 vm05 bash[20951]: cluster 2026-03-10T12:35:00.046409+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:01.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:01 vm05 bash[20951]: cluster 2026-03-10T12:35:00.046409+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:02.149 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:02.402 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:02.402 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"fsid":"f864ef90-1c7c-11f1-946a-0b343fff599a","created":"2026-03-10T12:32:36.746333+0000","modified":"2026-03-10T12:34:53.378039+0000","last_up_change":"2026-03-10T12:34:52.297933+0000","last_in_change":"2026-03-10T12:34:35.188063+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"dde3a957-3d01-45e3-a18f-a783cff4ef87","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6803","nonce":2723523180}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6805","nonce":2723523180}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6809","nonce":2723523180}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6807","nonce":2723523180}]},"public_addr":"192.168.123.101:6803/2723523180","cluster_addr":"192.168.123.101:6805/2723523180","heartbeat_back_addr":"192.168.123.101:6809/2723523180","heartbeat_front_addr":"192.168.123.101:6807/2723523180","state":["exists","up"]},{"osd":1,"uuid":"45acbc28-d814-49a2-a229-15ef9cf0619b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6801","nonce":3430972692}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6803","nonce":3430972692}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6807","nonce":3430972692}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6805","nonce":3430972692}]},"public_addr":"192.168.123.105:6801/3430972692","cluster_addr":"192.168.123.105:6803/3430972692","heartbeat_back_addr":"192.168.123.105:6807/3430972692","heartbeat_front_addr":"192.168.123.105:6805/3430972692","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:34:16.475588+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:34:49.497047+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:6800/2043439573":"2026-03-11T12:32:58.023984+0000","192.168.123.101:0/4145314724":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6801/2043439573":"2026-03-11T12:32:58.023984+0000","192.168.123.101:0/2638632070":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6801/3021433538":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/1624581309":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6800/3021433538":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/3204408413":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/648155406":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/3264267271":"2026-03-11T12:32:47.865414+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T12:35:02.451 INFO:tasks.cephadm.ceph_manager.ceph:[] 2026-03-10T12:35:02.451 INFO:tasks.cephadm:Setting up client nodes... 2026-03-10T12:35:02.451 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-10T12:35:02.451 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-10T12:35:02.451 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph mgr dump --format=json 2026-03-10T12:35:02.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:02 vm01 bash[20716]: audit 2026-03-10T12:35:02.401431+0000 mon.a (mon.0) 389 : audit [DBG] from='client.? 192.168.123.101:0/4256900742' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:02.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:02 vm01 bash[20716]: audit 2026-03-10T12:35:02.401431+0000 mon.a (mon.0) 389 : audit [DBG] from='client.? 192.168.123.101:0/4256900742' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:02.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:02 vm05 bash[20951]: audit 2026-03-10T12:35:02.401431+0000 mon.a (mon.0) 389 : audit [DBG] from='client.? 192.168.123.101:0/4256900742' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:02.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:02 vm05 bash[20951]: audit 2026-03-10T12:35:02.401431+0000 mon.a (mon.0) 389 : audit [DBG] from='client.? 192.168.123.101:0/4256900742' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:03.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:03 vm05 bash[20951]: cluster 2026-03-10T12:35:02.046625+0000 mgr.a (mgr.14150) 105 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:03.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:03 vm05 bash[20951]: cluster 2026-03-10T12:35:02.046625+0000 mgr.a (mgr.14150) 105 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:03.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:03 vm01 bash[20716]: cluster 2026-03-10T12:35:02.046625+0000 mgr.a (mgr.14150) 105 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:03.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:03 vm01 bash[20716]: cluster 2026-03-10T12:35:02.046625+0000 mgr.a (mgr.14150) 105 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:05.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:05 vm05 bash[20951]: cluster 2026-03-10T12:35:04.046843+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:05.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:05 vm05 bash[20951]: cluster 2026-03-10T12:35:04.046843+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:05.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:05 vm01 bash[20716]: cluster 2026-03-10T12:35:04.046843+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:05.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:05 vm01 bash[20716]: cluster 2026-03-10T12:35:04.046843+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:06.161 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:06.422 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:06.496 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"flags":0,"active_gid":14150,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":2589859639},{"type":"v1","addr":"192.168.123.101:6801","nonce":2589859639}]},"active_addr":"192.168.123.101:6801/2589859639","active_change":"2026-03-10T12:32:58.024235+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":24105,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.101:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3899335843}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":4286567742}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3530831896}]}]} 2026-03-10T12:35:06.497 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-10T12:35:06.497 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-10T12:35:06.497 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph osd dump --format=json 2026-03-10T12:35:06.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:06 vm01 bash[20716]: audit 2026-03-10T12:35:06.420488+0000 mon.a (mon.0) 390 : audit [DBG] from='client.? 192.168.123.101:0/3827765892' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T12:35:06.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:06 vm01 bash[20716]: audit 2026-03-10T12:35:06.420488+0000 mon.a (mon.0) 390 : audit [DBG] from='client.? 192.168.123.101:0/3827765892' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T12:35:06.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:06 vm05 bash[20951]: audit 2026-03-10T12:35:06.420488+0000 mon.a (mon.0) 390 : audit [DBG] from='client.? 192.168.123.101:0/3827765892' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T12:35:06.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:06 vm05 bash[20951]: audit 2026-03-10T12:35:06.420488+0000 mon.a (mon.0) 390 : audit [DBG] from='client.? 192.168.123.101:0/3827765892' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T12:35:07.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:07 vm05 bash[20951]: cluster 2026-03-10T12:35:06.047111+0000 mgr.a (mgr.14150) 107 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:07.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:07 vm05 bash[20951]: cluster 2026-03-10T12:35:06.047111+0000 mgr.a (mgr.14150) 107 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:07.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:07 vm01 bash[20716]: cluster 2026-03-10T12:35:06.047111+0000 mgr.a (mgr.14150) 107 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:07.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:07 vm01 bash[20716]: cluster 2026-03-10T12:35:06.047111+0000 mgr.a (mgr.14150) 107 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:09 vm05 bash[20951]: cluster 2026-03-10T12:35:08.047381+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:09 vm05 bash[20951]: cluster 2026-03-10T12:35:08.047381+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:09 vm01 bash[20716]: cluster 2026-03-10T12:35:08.047381+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:09 vm01 bash[20716]: cluster 2026-03-10T12:35:08.047381+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:10.174 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:10.420 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:10.420 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"fsid":"f864ef90-1c7c-11f1-946a-0b343fff599a","created":"2026-03-10T12:32:36.746333+0000","modified":"2026-03-10T12:34:53.378039+0000","last_up_change":"2026-03-10T12:34:52.297933+0000","last_in_change":"2026-03-10T12:34:35.188063+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"dde3a957-3d01-45e3-a18f-a783cff4ef87","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6803","nonce":2723523180}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6805","nonce":2723523180}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6809","nonce":2723523180}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6807","nonce":2723523180}]},"public_addr":"192.168.123.101:6803/2723523180","cluster_addr":"192.168.123.101:6805/2723523180","heartbeat_back_addr":"192.168.123.101:6809/2723523180","heartbeat_front_addr":"192.168.123.101:6807/2723523180","state":["exists","up"]},{"osd":1,"uuid":"45acbc28-d814-49a2-a229-15ef9cf0619b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6801","nonce":3430972692}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6803","nonce":3430972692}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6807","nonce":3430972692}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6805","nonce":3430972692}]},"public_addr":"192.168.123.105:6801/3430972692","cluster_addr":"192.168.123.105:6803/3430972692","heartbeat_back_addr":"192.168.123.105:6807/3430972692","heartbeat_front_addr":"192.168.123.105:6805/3430972692","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:34:16.475588+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:34:49.497047+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:6800/2043439573":"2026-03-11T12:32:58.023984+0000","192.168.123.101:0/4145314724":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6801/2043439573":"2026-03-11T12:32:58.023984+0000","192.168.123.101:0/2638632070":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6801/3021433538":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/1624581309":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6800/3021433538":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/3204408413":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/648155406":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/3264267271":"2026-03-11T12:32:47.865414+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T12:35:10.496 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-10T12:35:10.496 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph osd dump --format=json 2026-03-10T12:35:10.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:10 vm05 bash[20951]: audit 2026-03-10T12:35:10.419681+0000 mon.a (mon.0) 391 : audit [DBG] from='client.? 192.168.123.101:0/1648861153' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:10.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:10 vm05 bash[20951]: audit 2026-03-10T12:35:10.419681+0000 mon.a (mon.0) 391 : audit [DBG] from='client.? 192.168.123.101:0/1648861153' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:10.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:10 vm01 bash[20716]: audit 2026-03-10T12:35:10.419681+0000 mon.a (mon.0) 391 : audit [DBG] from='client.? 192.168.123.101:0/1648861153' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:10.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:10 vm01 bash[20716]: audit 2026-03-10T12:35:10.419681+0000 mon.a (mon.0) 391 : audit [DBG] from='client.? 192.168.123.101:0/1648861153' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:11 vm05 bash[20951]: cluster 2026-03-10T12:35:10.047624+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:11 vm05 bash[20951]: cluster 2026-03-10T12:35:10.047624+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:11 vm05 bash[20951]: audit 2026-03-10T12:35:11.282293+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:11 vm05 bash[20951]: audit 2026-03-10T12:35:11.282293+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:11 vm05 bash[20951]: audit 2026-03-10T12:35:11.286181+0000 mon.a (mon.0) 393 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:11 vm05 bash[20951]: audit 2026-03-10T12:35:11.286181+0000 mon.a (mon.0) 393 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:11 vm05 bash[20951]: audit 2026-03-10T12:35:11.290272+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:11 vm05 bash[20951]: audit 2026-03-10T12:35:11.290272+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:11 vm01 bash[20716]: cluster 2026-03-10T12:35:10.047624+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:11 vm01 bash[20716]: cluster 2026-03-10T12:35:10.047624+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:11 vm01 bash[20716]: audit 2026-03-10T12:35:11.282293+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:11 vm01 bash[20716]: audit 2026-03-10T12:35:11.282293+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:11 vm01 bash[20716]: audit 2026-03-10T12:35:11.286181+0000 mon.a (mon.0) 393 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:11 vm01 bash[20716]: audit 2026-03-10T12:35:11.286181+0000 mon.a (mon.0) 393 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:11 vm01 bash[20716]: audit 2026-03-10T12:35:11.290272+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:11 vm01 bash[20716]: audit 2026-03-10T12:35:11.290272+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:13 vm05 bash[20951]: cluster 2026-03-10T12:35:12.047879+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:14.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:13 vm05 bash[20951]: cluster 2026-03-10T12:35:12.047879+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:14.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:13 vm05 bash[20951]: audit 2026-03-10T12:35:12.770878+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:13 vm05 bash[20951]: audit 2026-03-10T12:35:12.770878+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:13 vm05 bash[20951]: audit 2026-03-10T12:35:12.779659+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:13 vm05 bash[20951]: audit 2026-03-10T12:35:12.779659+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:13 vm05 bash[20951]: audit 2026-03-10T12:35:12.799130+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:13 vm05 bash[20951]: audit 2026-03-10T12:35:12.799130+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.185 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:14.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:13 vm01 bash[20716]: cluster 2026-03-10T12:35:12.047879+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:14.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:13 vm01 bash[20716]: cluster 2026-03-10T12:35:12.047879+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:14.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:13 vm01 bash[20716]: audit 2026-03-10T12:35:12.770878+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:13 vm01 bash[20716]: audit 2026-03-10T12:35:12.770878+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:13 vm01 bash[20716]: audit 2026-03-10T12:35:12.779659+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:13 vm01 bash[20716]: audit 2026-03-10T12:35:12.779659+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:13 vm01 bash[20716]: audit 2026-03-10T12:35:12.799130+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:13 vm01 bash[20716]: audit 2026-03-10T12:35:12.799130+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:14.447 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:14.447 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"fsid":"f864ef90-1c7c-11f1-946a-0b343fff599a","created":"2026-03-10T12:32:36.746333+0000","modified":"2026-03-10T12:34:53.378039+0000","last_up_change":"2026-03-10T12:34:52.297933+0000","last_in_change":"2026-03-10T12:34:35.188063+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"dde3a957-3d01-45e3-a18f-a783cff4ef87","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6803","nonce":2723523180}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6805","nonce":2723523180}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6809","nonce":2723523180}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2723523180},{"type":"v1","addr":"192.168.123.101:6807","nonce":2723523180}]},"public_addr":"192.168.123.101:6803/2723523180","cluster_addr":"192.168.123.101:6805/2723523180","heartbeat_back_addr":"192.168.123.101:6809/2723523180","heartbeat_front_addr":"192.168.123.101:6807/2723523180","state":["exists","up"]},{"osd":1,"uuid":"45acbc28-d814-49a2-a229-15ef9cf0619b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6801","nonce":3430972692}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6803","nonce":3430972692}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6807","nonce":3430972692}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":3430972692},{"type":"v1","addr":"192.168.123.105:6805","nonce":3430972692}]},"public_addr":"192.168.123.105:6801/3430972692","cluster_addr":"192.168.123.105:6803/3430972692","heartbeat_back_addr":"192.168.123.105:6807/3430972692","heartbeat_front_addr":"192.168.123.105:6805/3430972692","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:34:16.475588+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:34:49.497047+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:6800/2043439573":"2026-03-11T12:32:58.023984+0000","192.168.123.101:0/4145314724":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6801/2043439573":"2026-03-11T12:32:58.023984+0000","192.168.123.101:0/2638632070":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6801/3021433538":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/1624581309":"2026-03-11T12:32:58.023984+0000","192.168.123.101:6800/3021433538":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/3204408413":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/648155406":"2026-03-11T12:32:47.865414+0000","192.168.123.101:0/3264267271":"2026-03-11T12:32:47.865414+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T12:35:14.500 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph tell osd.0 flush_pg_stats 2026-03-10T12:35:14.500 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph tell osd.1 flush_pg_stats 2026-03-10T12:35:14.775 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:14 vm01 bash[20716]: audit 2026-03-10T12:35:14.446781+0000 mon.a (mon.0) 398 : audit [DBG] from='client.? 192.168.123.101:0/2542636851' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:14.775 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:14 vm01 bash[20716]: audit 2026-03-10T12:35:14.446781+0000 mon.a (mon.0) 398 : audit [DBG] from='client.? 192.168.123.101:0/2542636851' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:15.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:14 vm05 bash[20951]: audit 2026-03-10T12:35:14.446781+0000 mon.a (mon.0) 398 : audit [DBG] from='client.? 192.168.123.101:0/2542636851' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:15.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:14 vm05 bash[20951]: audit 2026-03-10T12:35:14.446781+0000 mon.a (mon.0) 398 : audit [DBG] from='client.? 192.168.123.101:0/2542636851' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:35:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:15 vm05 bash[20951]: cluster 2026-03-10T12:35:14.048078+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:16.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:15 vm05 bash[20951]: cluster 2026-03-10T12:35:14.048078+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:16.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:15 vm01 bash[20716]: cluster 2026-03-10T12:35:14.048078+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:16.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:15 vm01 bash[20716]: cluster 2026-03-10T12:35:14.048078+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:18.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:17 vm05 bash[20951]: cluster 2026-03-10T12:35:16.048305+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:18.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:17 vm05 bash[20951]: cluster 2026-03-10T12:35:16.048305+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:18.198 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:18.200 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:18.201 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:17 vm01 bash[20716]: cluster 2026-03-10T12:35:16.048305+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:18.201 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:17 vm01 bash[20716]: cluster 2026-03-10T12:35:16.048305+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:18.610 INFO:teuthology.orchestra.run.vm01.stdout:55834574855 2026-03-10T12:35:18.610 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph osd last-stat-seq osd.1 2026-03-10T12:35:18.618 INFO:teuthology.orchestra.run.vm01.stdout:34359738381 2026-03-10T12:35:18.618 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph osd last-stat-seq osd.0 2026-03-10T12:35:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:19 vm01 bash[20716]: cluster 2026-03-10T12:35:18.048498+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:19 vm01 bash[20716]: cluster 2026-03-10T12:35:18.048498+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:20.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:19 vm05 bash[20951]: cluster 2026-03-10T12:35:18.048498+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:20.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:19 vm05 bash[20951]: cluster 2026-03-10T12:35:18.048498+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:22.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:21 vm05 bash[20951]: cluster 2026-03-10T12:35:20.048799+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:22.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:21 vm05 bash[20951]: cluster 2026-03-10T12:35:20.048799+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:22.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:21 vm01 bash[20716]: cluster 2026-03-10T12:35:20.048799+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:22.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:21 vm01 bash[20716]: cluster 2026-03-10T12:35:20.048799+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:23.231 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:23.232 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:23.490 INFO:teuthology.orchestra.run.vm01.stdout:34359738382 2026-03-10T12:35:23.517 INFO:teuthology.orchestra.run.vm01.stdout:55834574855 2026-03-10T12:35:23.559 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738381 got 34359738382 for osd.0 2026-03-10T12:35:23.559 DEBUG:teuthology.parallel:result is None 2026-03-10T12:35:23.591 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574855 got 55834574855 for osd.1 2026-03-10T12:35:23.591 DEBUG:teuthology.parallel:result is None 2026-03-10T12:35:23.591 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-10T12:35:23.591 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph pg dump --format=json 2026-03-10T12:35:23.803 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:23 vm01 bash[20716]: cluster 2026-03-10T12:35:22.049146+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:23.803 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:23 vm01 bash[20716]: cluster 2026-03-10T12:35:22.049146+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:23.803 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:23 vm01 bash[20716]: audit 2026-03-10T12:35:23.489263+0000 mon.a (mon.0) 399 : audit [DBG] from='client.? 192.168.123.101:0/1042505112' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:35:23.803 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:23 vm01 bash[20716]: audit 2026-03-10T12:35:23.489263+0000 mon.a (mon.0) 399 : audit [DBG] from='client.? 192.168.123.101:0/1042505112' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:35:23.803 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:23 vm01 bash[20716]: audit 2026-03-10T12:35:23.517181+0000 mon.b (mon.1) 12 : audit [DBG] from='client.? 192.168.123.101:0/3608202473' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:35:23.803 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:23 vm01 bash[20716]: audit 2026-03-10T12:35:23.517181+0000 mon.b (mon.1) 12 : audit [DBG] from='client.? 192.168.123.101:0/3608202473' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:35:24.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:23 vm05 bash[20951]: cluster 2026-03-10T12:35:22.049146+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:24.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:23 vm05 bash[20951]: cluster 2026-03-10T12:35:22.049146+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:24.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:23 vm05 bash[20951]: audit 2026-03-10T12:35:23.489263+0000 mon.a (mon.0) 399 : audit [DBG] from='client.? 192.168.123.101:0/1042505112' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:35:24.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:23 vm05 bash[20951]: audit 2026-03-10T12:35:23.489263+0000 mon.a (mon.0) 399 : audit [DBG] from='client.? 192.168.123.101:0/1042505112' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:35:24.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:23 vm05 bash[20951]: audit 2026-03-10T12:35:23.517181+0000 mon.b (mon.1) 12 : audit [DBG] from='client.? 192.168.123.101:0/3608202473' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:35:24.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:23 vm05 bash[20951]: audit 2026-03-10T12:35:23.517181+0000 mon.b (mon.1) 12 : audit [DBG] from='client.? 192.168.123.101:0/3608202473' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:35:26.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:25 vm05 bash[20951]: cluster 2026-03-10T12:35:24.049368+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:26.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:25 vm05 bash[20951]: cluster 2026-03-10T12:35:24.049368+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:26.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:25 vm01 bash[20716]: cluster 2026-03-10T12:35:24.049368+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:26.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:25 vm01 bash[20716]: cluster 2026-03-10T12:35:24.049368+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:27 vm05 bash[20951]: cluster 2026-03-10T12:35:26.049637+0000 mgr.a (mgr.14150) 117 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:28.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:27 vm05 bash[20951]: cluster 2026-03-10T12:35:26.049637+0000 mgr.a (mgr.14150) 117 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:28.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:27 vm01 bash[20716]: cluster 2026-03-10T12:35:26.049637+0000 mgr.a (mgr.14150) 117 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:28.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:27 vm01 bash[20716]: cluster 2026-03-10T12:35:26.049637+0000 mgr.a (mgr.14150) 117 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:28.242 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:28.481 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:28.481 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-10T12:35:28.526 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":79,"stamp":"2026-03-10T12:35:28.049808+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":53920,"kb_used_data":240,"kb_used_omap":3,"kb_used_meta":53628,"kb_avail":41880928,"statfs":{"total":42941284352,"available":42886070272,"internally_reserved":0,"allocated":245760,"data_stored":60148,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":3180,"internal_metadata":54915988},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":13,"seq":55834574857,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738383,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-10T12:35:28.527 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph pg dump --format=json 2026-03-10T12:35:30.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:29 vm01 bash[20716]: cluster 2026-03-10T12:35:28.049885+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:30.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:29 vm01 bash[20716]: cluster 2026-03-10T12:35:28.049885+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:30.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:29 vm01 bash[20716]: audit 2026-03-10T12:35:28.480273+0000 mgr.a (mgr.14150) 119 : audit [DBG] from='client.14270 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:30.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:29 vm01 bash[20716]: audit 2026-03-10T12:35:28.480273+0000 mgr.a (mgr.14150) 119 : audit [DBG] from='client.14270 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:30.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:29 vm05 bash[20951]: cluster 2026-03-10T12:35:28.049885+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:30.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:29 vm05 bash[20951]: cluster 2026-03-10T12:35:28.049885+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:30.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:29 vm05 bash[20951]: audit 2026-03-10T12:35:28.480273+0000 mgr.a (mgr.14150) 119 : audit [DBG] from='client.14270 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:30.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:29 vm05 bash[20951]: audit 2026-03-10T12:35:28.480273+0000 mgr.a (mgr.14150) 119 : audit [DBG] from='client.14270 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:31.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:30 vm01 bash[20716]: cluster 2026-03-10T12:35:30.050118+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:31.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:30 vm01 bash[20716]: cluster 2026-03-10T12:35:30.050118+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:31.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:30 vm05 bash[20951]: cluster 2026-03-10T12:35:30.050118+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:31.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:30 vm05 bash[20951]: cluster 2026-03-10T12:35:30.050118+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:32.261 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:32.507 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:32.507 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-10T12:35:32.524 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:32 vm01 bash[20716]: audit 2026-03-10T12:35:31.348000+0000 mon.a (mon.0) 400 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.524 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:32 vm01 bash[20716]: audit 2026-03-10T12:35:31.348000+0000 mon.a (mon.0) 400 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.524 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:32 vm01 bash[20716]: audit 2026-03-10T12:35:31.352428+0000 mon.a (mon.0) 401 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.524 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:32 vm01 bash[20716]: audit 2026-03-10T12:35:31.352428+0000 mon.a (mon.0) 401 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.524 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:32 vm01 bash[20716]: audit 2026-03-10T12:35:31.355994+0000 mon.a (mon.0) 402 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.524 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:32 vm01 bash[20716]: audit 2026-03-10T12:35:31.355994+0000 mon.a (mon.0) 402 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.559 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":81,"stamp":"2026-03-10T12:35:32.050300+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":53920,"kb_used_data":240,"kb_used_omap":3,"kb_used_meta":53628,"kb_avail":41880928,"statfs":{"total":42941284352,"available":42886070272,"internally_reserved":0,"allocated":245760,"data_stored":60148,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":3180,"internal_metadata":54915988},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":13,"seq":55834574857,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738384,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-10T12:35:32.559 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-10T12:35:32.559 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-10T12:35:32.559 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-10T12:35:32.559 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph health --format=json 2026-03-10T12:35:32.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:32 vm05 bash[20951]: audit 2026-03-10T12:35:31.348000+0000 mon.a (mon.0) 400 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:32 vm05 bash[20951]: audit 2026-03-10T12:35:31.348000+0000 mon.a (mon.0) 400 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:32 vm05 bash[20951]: audit 2026-03-10T12:35:31.352428+0000 mon.a (mon.0) 401 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:32 vm05 bash[20951]: audit 2026-03-10T12:35:31.352428+0000 mon.a (mon.0) 401 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:32 vm05 bash[20951]: audit 2026-03-10T12:35:31.355994+0000 mon.a (mon.0) 402 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:32.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:32 vm05 bash[20951]: audit 2026-03-10T12:35:31.355994+0000 mon.a (mon.0) 402 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: cluster 2026-03-10T12:35:32.050393+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:34.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: cluster 2026-03-10T12:35:32.050393+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:34.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: audit 2026-03-10T12:35:32.506614+0000 mgr.a (mgr.14150) 122 : audit [DBG] from='client.14274 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:34.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: audit 2026-03-10T12:35:32.506614+0000 mgr.a (mgr.14150) 122 : audit [DBG] from='client.14274 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:34.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: audit 2026-03-10T12:35:32.856379+0000 mon.a (mon.0) 403 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: audit 2026-03-10T12:35:32.856379+0000 mon.a (mon.0) 403 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: audit 2026-03-10T12:35:32.859086+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: audit 2026-03-10T12:35:32.859086+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: audit 2026-03-10T12:35:32.861989+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:33 vm05 bash[20951]: audit 2026-03-10T12:35:32.861989+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: cluster 2026-03-10T12:35:32.050393+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: cluster 2026-03-10T12:35:32.050393+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: audit 2026-03-10T12:35:32.506614+0000 mgr.a (mgr.14150) 122 : audit [DBG] from='client.14274 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: audit 2026-03-10T12:35:32.506614+0000 mgr.a (mgr.14150) 122 : audit [DBG] from='client.14274 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: audit 2026-03-10T12:35:32.856379+0000 mon.a (mon.0) 403 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: audit 2026-03-10T12:35:32.856379+0000 mon.a (mon.0) 403 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: audit 2026-03-10T12:35:32.859086+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: audit 2026-03-10T12:35:32.859086+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: audit 2026-03-10T12:35:32.861989+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:34.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:33 vm01 bash[20716]: audit 2026-03-10T12:35:32.861989+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:36.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:35 vm05 bash[20951]: cluster 2026-03-10T12:35:34.050585+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:36.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:35 vm05 bash[20951]: cluster 2026-03-10T12:35:34.050585+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:36.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:35 vm01 bash[20716]: cluster 2026-03-10T12:35:34.050585+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:36.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:35 vm01 bash[20716]: cluster 2026-03-10T12:35:34.050585+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:36.271 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:36.549 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:36.549 INFO:teuthology.orchestra.run.vm01.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-10T12:35:36.715 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-10T12:35:36.715 INFO:tasks.cephadm:Setup complete, yielding 2026-03-10T12:35:36.715 INFO:teuthology.run_tasks:Running task exec... 2026-03-10T12:35:36.718 INFO:teuthology.task.exec:Executing custom commands... 2026-03-10T12:35:36.718 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T12:35:36.718 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-10T12:35:36.726 INFO:teuthology.task.exec:Running commands on role host.b host ubuntu@vm05.local 2026-03-10T12:35:36.726 DEBUG:teuthology.orchestra.run.vm05:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-10T12:35:36.736 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-10T12:35:36.739 INFO:tasks.cephadm:Applying spec(s): extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon --- custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest --- custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z 2026-03-10T12:35:36.739 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch apply -i - 2026-03-10T12:35:37.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:36 vm05 bash[20951]: audit 2026-03-10T12:35:36.548886+0000 mon.a (mon.0) 406 : audit [DBG] from='client.? 192.168.123.101:0/141261964' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T12:35:37.175 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:36 vm05 bash[20951]: audit 2026-03-10T12:35:36.548886+0000 mon.a (mon.0) 406 : audit [DBG] from='client.? 192.168.123.101:0/141261964' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T12:35:37.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:36 vm01 bash[20716]: audit 2026-03-10T12:35:36.548886+0000 mon.a (mon.0) 406 : audit [DBG] from='client.? 192.168.123.101:0/141261964' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T12:35:37.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:36 vm01 bash[20716]: audit 2026-03-10T12:35:36.548886+0000 mon.a (mon.0) 406 : audit [DBG] from='client.? 192.168.123.101:0/141261964' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T12:35:38.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:37 vm05 bash[20951]: cluster 2026-03-10T12:35:36.050803+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:38.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:37 vm05 bash[20951]: cluster 2026-03-10T12:35:36.050803+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:38.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:37 vm01 bash[20716]: cluster 2026-03-10T12:35:36.050803+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:38.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:37 vm01 bash[20716]: cluster 2026-03-10T12:35:36.050803+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:40.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:39 vm05 bash[20951]: cluster 2026-03-10T12:35:38.051033+0000 mgr.a (mgr.14150) 125 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:40.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:39 vm05 bash[20951]: cluster 2026-03-10T12:35:38.051033+0000 mgr.a (mgr.14150) 125 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:40.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:39 vm01 bash[20716]: cluster 2026-03-10T12:35:38.051033+0000 mgr.a (mgr.14150) 125 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:40.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:39 vm01 bash[20716]: cluster 2026-03-10T12:35:38.051033+0000 mgr.a (mgr.14150) 125 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:41.251 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:40 vm01 bash[20716]: cluster 2026-03-10T12:35:40.051250+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:41.251 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:40 vm01 bash[20716]: cluster 2026-03-10T12:35:40.051250+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:41.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:40 vm05 bash[20951]: cluster 2026-03-10T12:35:40.051250+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:41.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:40 vm05 bash[20951]: cluster 2026-03-10T12:35:40.051250+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:41.775 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:42.059 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled mon update... 2026-03-10T12:35:42.059 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled container.foo update... 2026-03-10T12:35:42.059 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled container.bar update... 2026-03-10T12:35:42.147 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-10T12:35:42.149 INFO:tasks.cephadm:Waiting for ceph service mon to start (timeout 300)... 2026-03-10T12:35:42.150 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch ls -f json 2026-03-10T12:35:42.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:42 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:42.699 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:35:42 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:42.699 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:35:42 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.051 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:35:42 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.051 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:42 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.051 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:35:42 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.402 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.040770+0000 mgr.a (mgr.14150) 127 : audit [DBG] from='client.24163 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.040770+0000 mgr.a (mgr.14150) 127 : audit [DBG] from='client.24163 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.045849+0000 mgr.a (mgr.14150) 128 : cephadm [INF] Saving service mon spec with placement * 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.045849+0000 mgr.a (mgr.14150) 128 : cephadm [INF] Saving service mon spec with placement * 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.050103+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.050103+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.050846+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.050846+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.050969+0000 mgr.a (mgr.14150) 129 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.050969+0000 mgr.a (mgr.14150) 129 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cluster 2026-03-10T12:35:42.051502+0000 mgr.a (mgr.14150) 130 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cluster 2026-03-10T12:35:42.051502+0000 mgr.a (mgr.14150) 130 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.052762+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.052762+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.053235+0000 mon.a (mon.0) 410 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.053235+0000 mon.a (mon.0) 410 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.054570+0000 mon.a (mon.0) 411 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.054570+0000 mon.a (mon.0) 411 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.054942+0000 mgr.a (mgr.14150) 131 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.054942+0000 mgr.a (mgr.14150) 131 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.058231+0000 mon.a (mon.0) 412 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.058231+0000 mon.a (mon.0) 412 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.062912+0000 mon.a (mon.0) 413 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.062912+0000 mon.a (mon.0) 413 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.069178+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.069178+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.070272+0000 mgr.a (mgr.14150) 132 : cephadm [INF] Deploying daemon container.foo.vm01 on vm01 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.070272+0000 mgr.a (mgr.14150) 132 : cephadm [INF] Deploying daemon container.foo.vm01 on vm01 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.824936+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.824936+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.829146+0000 mon.a (mon.0) 416 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.829146+0000 mon.a (mon.0) 416 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.832149+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: audit 2026-03-10T12:35:42.832149+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.832746+0000 mgr.a (mgr.14150) 133 : cephadm [INF] Deploying daemon container.foo.vm05 on vm05 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 bash[20951]: cephadm 2026-03-10T12:35:42.832746+0000 mgr.a (mgr.14150) 133 : cephadm [INF] Deploying daemon container.foo.vm05 on vm05 2026-03-10T12:35:43.403 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.403 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.040770+0000 mgr.a (mgr.14150) 127 : audit [DBG] from='client.24163 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.040770+0000 mgr.a (mgr.14150) 127 : audit [DBG] from='client.24163 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.045849+0000 mgr.a (mgr.14150) 128 : cephadm [INF] Saving service mon spec with placement * 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.045849+0000 mgr.a (mgr.14150) 128 : cephadm [INF] Saving service mon spec with placement * 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.050103+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.050103+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.050846+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.050846+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.050969+0000 mgr.a (mgr.14150) 129 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-10T12:35:43.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.050969+0000 mgr.a (mgr.14150) 129 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cluster 2026-03-10T12:35:42.051502+0000 mgr.a (mgr.14150) 130 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cluster 2026-03-10T12:35:42.051502+0000 mgr.a (mgr.14150) 130 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.052762+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.052762+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.053235+0000 mon.a (mon.0) 410 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.053235+0000 mon.a (mon.0) 410 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.054570+0000 mon.a (mon.0) 411 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.054570+0000 mon.a (mon.0) 411 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.054942+0000 mgr.a (mgr.14150) 131 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.054942+0000 mgr.a (mgr.14150) 131 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.058231+0000 mon.a (mon.0) 412 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.058231+0000 mon.a (mon.0) 412 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.062912+0000 mon.a (mon.0) 413 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.062912+0000 mon.a (mon.0) 413 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.069178+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.069178+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.070272+0000 mgr.a (mgr.14150) 132 : cephadm [INF] Deploying daemon container.foo.vm01 on vm01 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.070272+0000 mgr.a (mgr.14150) 132 : cephadm [INF] Deploying daemon container.foo.vm01 on vm01 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.824936+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.824936+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.829146+0000 mon.a (mon.0) 416 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.829146+0000 mon.a (mon.0) 416 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.832149+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: audit 2026-03-10T12:35:42.832149+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.832746+0000 mgr.a (mgr.14150) 133 : cephadm [INF] Deploying daemon container.foo.vm05 on vm05 2026-03-10T12:35:43.450 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:43 vm01 bash[20716]: cephadm 2026-03-10T12:35:42.832746+0000 mgr.a (mgr.14150) 133 : cephadm [INF] Deploying daemon container.foo.vm05 on vm05 2026-03-10T12:35:43.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.674 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:43.674 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:44.161 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:44.162 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:44.162 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:35:43 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:44.424 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:35:44 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:44.424 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:35:44 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:44.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: audit 2026-03-10T12:35:43.506790+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: audit 2026-03-10T12:35:43.506790+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: audit 2026-03-10T12:35:43.510377+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: audit 2026-03-10T12:35:43.510377+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: audit 2026-03-10T12:35:43.514042+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: audit 2026-03-10T12:35:43.514042+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: audit 2026-03-10T12:35:43.517018+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: audit 2026-03-10T12:35:43.517018+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: cephadm 2026-03-10T12:35:43.520299+0000 mgr.a (mgr.14150) 134 : cephadm [INF] Deploying daemon container.bar.vm05 on vm05 2026-03-10T12:35:44.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:44 vm01 bash[20716]: cephadm 2026-03-10T12:35:43.520299+0000 mgr.a (mgr.14150) 134 : cephadm [INF] Deploying daemon container.bar.vm05 on vm05 2026-03-10T12:35:44.887 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: audit 2026-03-10T12:35:43.506790+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.887 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: audit 2026-03-10T12:35:43.506790+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.887 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: audit 2026-03-10T12:35:43.510377+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.887 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: audit 2026-03-10T12:35:43.510377+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.887 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: audit 2026-03-10T12:35:43.514042+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.887 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: audit 2026-03-10T12:35:43.514042+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.887 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: audit 2026-03-10T12:35:43.517018+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.887 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: audit 2026-03-10T12:35:43.517018+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:44.888 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: cephadm 2026-03-10T12:35:43.520299+0000 mgr.a (mgr.14150) 134 : cephadm [INF] Deploying daemon container.bar.vm05 on vm05 2026-03-10T12:35:44.888 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:44 vm05 bash[20951]: cephadm 2026-03-10T12:35:43.520299+0000 mgr.a (mgr.14150) 134 : cephadm [INF] Deploying daemon container.bar.vm05 on vm05 2026-03-10T12:35:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:45 vm05 bash[20951]: cluster 2026-03-10T12:35:44.051702+0000 mgr.a (mgr.14150) 135 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:45 vm05 bash[20951]: cluster 2026-03-10T12:35:44.051702+0000 mgr.a (mgr.14150) 135 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:45 vm05 bash[20951]: audit 2026-03-10T12:35:44.894156+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:45 vm05 bash[20951]: audit 2026-03-10T12:35:44.894156+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:45 vm05 bash[20951]: audit 2026-03-10T12:35:45.190745+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:45.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:45 vm05 bash[20951]: audit 2026-03-10T12:35:45.190745+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:46.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:45 vm01 bash[20716]: cluster 2026-03-10T12:35:44.051702+0000 mgr.a (mgr.14150) 135 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:46.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:45 vm01 bash[20716]: cluster 2026-03-10T12:35:44.051702+0000 mgr.a (mgr.14150) 135 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:46.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:45 vm01 bash[20716]: audit 2026-03-10T12:35:44.894156+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:46.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:45 vm01 bash[20716]: audit 2026-03-10T12:35:44.894156+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:46.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:45 vm01 bash[20716]: audit 2026-03-10T12:35:45.190745+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:46.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:45 vm01 bash[20716]: audit 2026-03-10T12:35:45.190745+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:46.768 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:47.140 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:47.140 INFO:teuthology.orchestra.run.vm01.stdout:[{"events": ["2026-03-10T12:33:19.615978Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-10T12:32:58.048040Z", "last_refresh": "2026-03-10T12:35:31.342702Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-10T12:35:42.058523Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-10T12:35:42.054947Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-10T12:35:43.517376Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-10T12:35:42.050980Z", "running": 0, "size": 2}}, {"events": ["2026-03-10T12:33:46.420849Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm01=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T12:33:45.549327Z", "last_refresh": "2026-03-10T12:35:31.342667Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:35:42.069424Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T12:35:42.045863Z", "last_refresh": "2026-03-10T12:35:31.342775Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-10T12:35:31.342739Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-10T12:35:47.210 INFO:tasks.cephadm:mon has 2/2 2026-03-10T12:35:47.210 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-10T12:35:47.212 INFO:tasks.cephadm:Waiting for ceph service container.foo to start (timeout 300)... 2026-03-10T12:35:47.212 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch ls -f json 2026-03-10T12:35:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:47 vm01 bash[20716]: cluster 2026-03-10T12:35:46.051966+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:48.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:47 vm01 bash[20716]: cluster 2026-03-10T12:35:46.051966+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:48.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:47 vm05 bash[20951]: cluster 2026-03-10T12:35:46.051966+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:48.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:47 vm05 bash[20951]: cluster 2026-03-10T12:35:46.051966+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:49.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:49 vm05 bash[20951]: audit 2026-03-10T12:35:47.138731+0000 mgr.a (mgr.14150) 137 : audit [DBG] from='client.24165 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:49.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:49 vm05 bash[20951]: audit 2026-03-10T12:35:47.138731+0000 mgr.a (mgr.14150) 137 : audit [DBG] from='client.24165 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:49.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:49 vm05 bash[20951]: cluster 2026-03-10T12:35:48.052190+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:49.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:49 vm05 bash[20951]: cluster 2026-03-10T12:35:48.052190+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:49.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:49 vm01 bash[20716]: audit 2026-03-10T12:35:47.138731+0000 mgr.a (mgr.14150) 137 : audit [DBG] from='client.24165 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:49.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:49 vm01 bash[20716]: audit 2026-03-10T12:35:47.138731+0000 mgr.a (mgr.14150) 137 : audit [DBG] from='client.24165 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:49.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:49 vm01 bash[20716]: cluster 2026-03-10T12:35:48.052190+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:49.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:49 vm01 bash[20716]: cluster 2026-03-10T12:35:48.052190+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:51.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:51 vm05 bash[20951]: cluster 2026-03-10T12:35:50.052455+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:51.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:51 vm05 bash[20951]: cluster 2026-03-10T12:35:50.052455+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:51.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:51 vm05 bash[20951]: audit 2026-03-10T12:35:51.096053+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:51.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:51 vm05 bash[20951]: audit 2026-03-10T12:35:51.096053+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:51.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:51 vm01 bash[20716]: cluster 2026-03-10T12:35:50.052455+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:51.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:51 vm01 bash[20716]: cluster 2026-03-10T12:35:50.052455+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:51.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:51 vm01 bash[20716]: audit 2026-03-10T12:35:51.096053+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:51.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:51 vm01 bash[20716]: audit 2026-03-10T12:35:51.096053+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.300 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:52 vm05 bash[20951]: audit 2026-03-10T12:35:51.200809+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.300 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:52 vm05 bash[20951]: audit 2026-03-10T12:35:51.200809+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.300 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:52 vm05 bash[20951]: audit 2026-03-10T12:35:51.210804+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.300 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:52 vm05 bash[20951]: audit 2026-03-10T12:35:51.210804+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:52 vm01 bash[20716]: audit 2026-03-10T12:35:51.200809+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:52 vm01 bash[20716]: audit 2026-03-10T12:35:51.200809+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:52 vm01 bash[20716]: audit 2026-03-10T12:35:51.210804+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:52 vm01 bash[20716]: audit 2026-03-10T12:35:51.210804+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:52.785 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:53.055 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:53.056 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:35:53 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:53.056 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:35:53 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:53.265 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:53.265 INFO:teuthology.orchestra.run.vm01.stdout:[{"events": ["2026-03-10T12:33:19.615978Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-10T12:32:58.048040Z", "last_refresh": "2026-03-10T12:35:51.024563Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-10T12:35:52.558691Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-10T12:35:42.054947Z", "last_refresh": "2026-03-10T12:35:53.144992Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-10T12:35:43.517376Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-10T12:35:42.050980Z", "last_refresh": "2026-03-10T12:35:51.024486Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:33:46.420849Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm01=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T12:33:45.549327Z", "last_refresh": "2026-03-10T12:35:51.024543Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:35:42.069424Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T12:35:42.045863Z", "last_refresh": "2026-03-10T12:35:51.024620Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-10T12:35:51.024592Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-10T12:35:53.319 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:53.319 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:35:53 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:53.319 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:35:53 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:35:53.320 INFO:tasks.cephadm:container.foo has 2/2 2026-03-10T12:35:53.320 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-10T12:35:53.322 INFO:tasks.cephadm:Waiting for ceph service container.bar to start (timeout 300)... 2026-03-10T12:35:53.322 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch ls -f json 2026-03-10T12:35:53.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: cluster 2026-03-10T12:35:52.052750+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:53.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: cluster 2026-03-10T12:35:52.052750+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:53.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.326833+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.326833+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.336239+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.336239+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.341729+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.341729+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.551656+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.551656+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.555022+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.555022+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.558472+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:52.558472+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: cephadm 2026-03-10T12:35:52.559113+0000 mgr.a (mgr.14150) 141 : cephadm [INF] Deploying daemon container.bar.vm01 on vm01 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: cephadm 2026-03-10T12:35:52.559113+0000 mgr.a (mgr.14150) 141 : cephadm [INF] Deploying daemon container.bar.vm01 on vm01 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:53.151758+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:53.151758+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:53.159401+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:53.159401+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:53.166001+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:53 vm05 bash[20951]: audit 2026-03-10T12:35:53.166001+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: cluster 2026-03-10T12:35:52.052750+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: cluster 2026-03-10T12:35:52.052750+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.326833+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.326833+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.336239+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.336239+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.341729+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.341729+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.551656+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.551656+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.555022+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.555022+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.558472+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:52.558472+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: cephadm 2026-03-10T12:35:52.559113+0000 mgr.a (mgr.14150) 141 : cephadm [INF] Deploying daemon container.bar.vm01 on vm01 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: cephadm 2026-03-10T12:35:52.559113+0000 mgr.a (mgr.14150) 141 : cephadm [INF] Deploying daemon container.bar.vm01 on vm01 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:53.151758+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:53.151758+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:53.159401+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:53.159401+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:53.166001+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:53.700 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:53 vm01 bash[20716]: audit 2026-03-10T12:35:53.166001+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:35:54.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:54 vm05 bash[20951]: audit 2026-03-10T12:35:53.263741+0000 mgr.a (mgr.14150) 142 : audit [DBG] from='client.14290 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:54.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:54 vm05 bash[20951]: audit 2026-03-10T12:35:53.263741+0000 mgr.a (mgr.14150) 142 : audit [DBG] from='client.14290 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:54 vm01 bash[20716]: audit 2026-03-10T12:35:53.263741+0000 mgr.a (mgr.14150) 142 : audit [DBG] from='client.14290 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:54.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:54 vm01 bash[20716]: audit 2026-03-10T12:35:53.263741+0000 mgr.a (mgr.14150) 142 : audit [DBG] from='client.14290 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:55.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:55 vm01 bash[20716]: cluster 2026-03-10T12:35:54.052981+0000 mgr.a (mgr.14150) 143 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:55.449 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:55 vm01 bash[20716]: cluster 2026-03-10T12:35:54.052981+0000 mgr.a (mgr.14150) 143 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:55.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:55 vm05 bash[20951]: cluster 2026-03-10T12:35:54.052981+0000 mgr.a (mgr.14150) 143 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:55.675 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:55 vm05 bash[20951]: cluster 2026-03-10T12:35:54.052981+0000 mgr.a (mgr.14150) 143 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:57.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:57 vm05 bash[20951]: cluster 2026-03-10T12:35:56.053241+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:57.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:57 vm05 bash[20951]: cluster 2026-03-10T12:35:56.053241+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:57.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:57 vm01 bash[20716]: cluster 2026-03-10T12:35:56.053241+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:57.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:57 vm01 bash[20716]: cluster 2026-03-10T12:35:56.053241+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:57.934 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:35:58.343 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:35:58.343 INFO:teuthology.orchestra.run.vm01.stdout:[{"events": ["2026-03-10T12:33:19.615978Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-10T12:32:58.048040Z", "last_refresh": "2026-03-10T12:35:51.024563Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-10T12:35:52.558691Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-10T12:35:42.054947Z", "last_refresh": "2026-03-10T12:35:53.144992Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-10T12:35:43.517376Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-10T12:35:42.050980Z", "last_refresh": "2026-03-10T12:35:51.024486Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:33:46.420849Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm01=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T12:33:45.549327Z", "last_refresh": "2026-03-10T12:35:51.024543Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:35:42.069424Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T12:35:42.045863Z", "last_refresh": "2026-03-10T12:35:51.024620Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-10T12:35:51.024592Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-10T12:35:58.418 INFO:tasks.cephadm:container.bar has 0/2 2026-03-10T12:35:59.419 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch ls -f json 2026-03-10T12:35:59.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:59 vm05 bash[20951]: cluster 2026-03-10T12:35:58.053459+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v94: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:59.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:59 vm05 bash[20951]: cluster 2026-03-10T12:35:58.053459+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v94: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:59.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:59 vm05 bash[20951]: audit 2026-03-10T12:35:58.342116+0000 mgr.a (mgr.14150) 146 : audit [DBG] from='client.14294 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:59.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:35:59 vm05 bash[20951]: audit 2026-03-10T12:35:58.342116+0000 mgr.a (mgr.14150) 146 : audit [DBG] from='client.14294 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:59.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:59 vm01 bash[20716]: cluster 2026-03-10T12:35:58.053459+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v94: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:59.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:59 vm01 bash[20716]: cluster 2026-03-10T12:35:58.053459+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v94: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:35:59.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:59 vm01 bash[20716]: audit 2026-03-10T12:35:58.342116+0000 mgr.a (mgr.14150) 146 : audit [DBG] from='client.14294 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:35:59.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:35:59 vm01 bash[20716]: audit 2026-03-10T12:35:58.342116+0000 mgr.a (mgr.14150) 146 : audit [DBG] from='client.14294 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:01.925 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:01 vm01 bash[20716]: cluster 2026-03-10T12:36:00.053697+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v95: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:01.926 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:01 vm01 bash[20716]: cluster 2026-03-10T12:36:00.053697+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v95: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:02.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:01 vm05 bash[20951]: cluster 2026-03-10T12:36:00.053697+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v95: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:02.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:01 vm05 bash[20951]: cluster 2026-03-10T12:36:00.053697+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v95: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:02.900 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:02 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:02.900 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:02 vm01 systemd[1]: Stopping Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:36:02.900 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:02 vm01 bash[20716]: debug 2026-03-10T12:36:02.744+0000 7f32c6a89640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-10T12:36:02.900 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:02 vm01 bash[20716]: debug 2026-03-10T12:36:02.744+0000 7f32c6a89640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-10T12:36:02.900 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:36:02 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:02.900 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:36:02 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:03.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:02 vm01 bash[34281]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-a 2026-03-10T12:36:03.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:02 vm01 systemd[1]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a.service: Deactivated successfully. 2026-03-10T12:36:03.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:02 vm01 systemd[1]: Stopped Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:36:03.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:03.199 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:36:03 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:03.199 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:36:03 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 systemd[1]: Started Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.308+0000 7f2ed5dc0d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.308+0000 7f2ed5dc0d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.308+0000 7f2ed5dc0d80 0 pidfile_write: ignore empty --pid-file 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 0 load: jerasure load: lrc 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Git sha 0 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: DB SUMMARY 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: DB Session ID: J8DVZH05EGYB772TPUWO 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: MANIFEST file: MANIFEST-000015 size: 281 Bytes 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000014.log size: 4599502 ; 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T12:36:03.607 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.env: 0x55978c17cdc0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.info_log: 0x5597b8fd8700 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.statistics: (nil) 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.use_fsync: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.db_log_dir: 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.wal_dir: 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.write_buffer_manager: 0x5597b8fdd900 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.unordered_write: 0 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.row_cache: None 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.wal_filter: None 2026-03-10T12:36:03.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.wal_compression: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_open_files: -1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Compression algorithms supported: 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: kZSTD supported: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.merge_operator: 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_filter: None 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5597b8fd8640) 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cache_index_and_filter_blocks: 1 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T12:36:03.609 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: pin_top_level_index_and_filter: 1 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: index_type: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: data_block_index_type: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: index_shortening: 1 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: checksum: 4 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: no_block_cache: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: block_cache: 0x5597b8fff350 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: block_cache_name: BinnedLRUCache 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: block_cache_options: 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: capacity : 536870912 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: num_shard_bits : 4 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: strict_capacity_limit : 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: high_pri_pool_ratio: 0.000 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: block_cache_compressed: (nil) 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: persistent_cache: (nil) 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: block_size: 4096 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: block_size_deviation: 10 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: block_restart_interval: 16 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: index_block_restart_interval: 1 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: metadata_block_size: 4096 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: partition_filters: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: use_delta_encoding: 1 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: filter_policy: bloomfilter 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: whole_key_filtering: 1 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: verify_compression: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: read_amp_bytes_per_bit: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: format_version: 5 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: enable_index_compression: 1 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: block_align: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: max_auto_readahead_size: 262144 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: prepopulate_block_cache: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: initial_auto_readahead_size: 8192 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: num_file_reads_for_auto_readahead: 2 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression: NoCompression 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.num_levels: 7 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:36:03.610 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.ttl: 2592000 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 225, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 38f355c1-e07f-4226-a178-26f487d5fa48 2026-03-10T12:36:03.611 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146163318204, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.312+0000 7f2ed5dc0d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.332+0000 7f2ed5dc0d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146163335442, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 3954313, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 230, "largest_seqno": 3867, "table_properties": {"data_size": 3941106, "index_size": 8198, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3973, "raw_key_size": 34704, "raw_average_key_size": 21, "raw_value_size": 3910507, "raw_average_value_size": 2478, "num_data_blocks": 392, "num_entries": 1578, "num_filter_entries": 1578, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773146163, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "38f355c1-e07f-4226-a178-26f487d5fa48", "db_session_id": "J8DVZH05EGYB772TPUWO", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.332+0000 7f2ed5dc0d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146163335565, "job": 1, "event": "recovery_finished"} 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.332+0000 7f2ed5dc0d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.332+0000 7f2ed5dc0d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.332+0000 7f2ed5dc0d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5597b9000e00 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 4 rocksdb: DB pointer 0x5597b9df2000 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] at bind addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 1 mon.a@-1(???) e2 preinit fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 0 mon.a@-1(???).mds e1 new map 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 0 mon.a@-1(???).mds e1 print_map 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: e1 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: btime 2026-03-10T12:32:36:745933+0000 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: legacy client fscid: -1 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: No filesystems configured 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 0 mon.a@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:03.612 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: debug 2026-03-10T12:36:03.336+0000 7f2ed5dc0d80 1 mon.a@-1(???).paxosservice(auth 1..9) refresh upgraded, format 0 -> 3 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:02.759860+0000 mon.b (mon.1) 13 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:02.759860+0000 mon.b (mon.1) 13 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:02.760452+0000 mon.b (mon.1) 14 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:02.760452+0000 mon.b (mon.1) 14 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.547651+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.547651+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.550232+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.550232+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553153+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553153+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553173+0000 mon.a (mon.0) 4 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553173+0000 mon.a (mon.0) 4 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553182+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553182+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553192+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553192+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553200+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553200+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553208+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553208+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553218+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553218+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553227+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553227+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553554+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553554+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553574+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.553574+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.554224+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.554224+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.554315+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: cluster 2026-03-10T12:36:03.554315+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.561180+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.561180+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:03.871 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.565440+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.565440+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.571483+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.571483+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.575597+0000 mon.b (mon.1) 15 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.575597+0000 mon.b (mon.1) 15 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.577800+0000 mon.b (mon.1) 16 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.577800+0000 mon.b (mon.1) 16 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.578654+0000 mon.b (mon.1) 17 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:03.872 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:03 vm05 bash[20951]: audit 2026-03-10T12:36:03.578654+0000 mon.b (mon.1) 17 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:02.759860+0000 mon.b (mon.1) 13 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:02.759860+0000 mon.b (mon.1) 13 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:02.760452+0000 mon.b (mon.1) 14 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:02.760452+0000 mon.b (mon.1) 14 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.547651+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.547651+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.550232+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.550232+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:03.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553153+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553153+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553173+0000 mon.a (mon.0) 4 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553173+0000 mon.a (mon.0) 4 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553182+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553182+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553192+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553192+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553200+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553200+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553208+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553208+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553218+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553218+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553227+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553227+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553554+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553554+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553574+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.553574+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.554224+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.554224+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.554315+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: cluster 2026-03-10T12:36:03.554315+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.561180+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.561180+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.565440+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.565440+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.571483+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.571483+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.575597+0000 mon.b (mon.1) 15 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.575597+0000 mon.b (mon.1) 15 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.577800+0000 mon.b (mon.1) 16 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.577800+0000 mon.b (mon.1) 16 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.578654+0000 mon.b (mon.1) 17 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:03.940 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:03 vm01 bash[34421]: audit 2026-03-10T12:36:03.578654+0000 mon.b (mon.1) 17 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:04.158 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:04.158 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: Stopping Ceph mon.b for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:36:04.158 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:04.158 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:04.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[20951]: debug 2026-03-10T12:36:04.195+0000 7fb3471f0640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-10T12:36:04.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[20951]: debug 2026-03-10T12:36:04.195+0000 7fb3471f0640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-10T12:36:04.704 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[28916]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-b 2026-03-10T12:36:04.704 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.b.service: Deactivated successfully. 2026-03-10T12:36:04.704 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: Stopped Ceph mon.b for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:36:04.704 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:04.705 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:04.705 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:05.085 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 systemd[1]: Started Ceph mon.b for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.819+0000 7fe4ececbd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.819+0000 7fe4ececbd80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.819+0000 7fe4ececbd80 0 pidfile_write: ignore empty --pid-file 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 0 load: jerasure load: lrc 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Git sha 0 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: DB SUMMARY 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: DB Session ID: SNEVHB23H1KSZW4FJ15T 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T12:36:05.163 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 1, files: 000008.sst 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000009.log size: 5966202 ; 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.env: 0x55fbf27f6dc0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.info_log: 0x55fc050ecd00 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.statistics: (nil) 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.use_fsync: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.db_log_dir: 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.wal_dir: 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.write_buffer_manager: 0x55fc050f1900 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.unordered_write: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.row_cache: None 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.wal_filter: None 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.wal_compression: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T12:36:05.164 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_open_files: -1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Compression algorithms supported: 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: kZSTD supported: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.823+0000 7fe4ececbd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.merge_operator: 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_filter: None 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55fc050ec480) 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: cache_index_and_filter_blocks: 1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: pin_top_level_index_and_filter: 1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: index_type: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: data_block_index_type: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: index_shortening: 1 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: checksum: 4 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: no_block_cache: 0 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: block_cache: 0x55fc05113350 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: block_cache_name: BinnedLRUCache 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: block_cache_options: 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: capacity : 536870912 2026-03-10T12:36:05.165 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: num_shard_bits : 4 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: strict_capacity_limit : 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: high_pri_pool_ratio: 0.000 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: block_cache_compressed: (nil) 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: persistent_cache: (nil) 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: block_size: 4096 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: block_size_deviation: 10 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: block_restart_interval: 16 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: index_block_restart_interval: 1 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: metadata_block_size: 4096 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: partition_filters: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: use_delta_encoding: 1 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: filter_policy: bloomfilter 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: whole_key_filtering: 1 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: verify_compression: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: read_amp_bytes_per_bit: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: format_version: 5 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: enable_index_compression: 1 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: block_align: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: max_auto_readahead_size: 262144 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: prepopulate_block_cache: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: initial_auto_readahead_size: 8192 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: num_file_reads_for_auto_readahead: 2 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression: NoCompression 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.num_levels: 7 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T12:36:05.166 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.ttl: 2592000 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T12:36:05.167 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.827+0000 7fe4ececbd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.831+0000 7fe4ececbd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.831+0000 7fe4ececbd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.831+0000 7fe4ececbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 6bd5b8b6-8815-4e14-b03d-5213c86c0db1 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.831+0000 7fe4ececbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146164835212, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.831+0000 7fe4ececbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.851+0000 7fe4ececbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146164854756, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 4080061, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 6, "largest_seqno": 3831, "table_properties": {"data_size": 4065977, "index_size": 8563, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 4485, "raw_key_size": 44651, "raw_average_key_size": 25, "raw_value_size": 4029784, "raw_average_value_size": 2256, "num_data_blocks": 406, "num_entries": 1786, "num_filter_entries": 1786, "num_deletions": 4, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773146164, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "6bd5b8b6-8815-4e14-b03d-5213c86c0db1", "db_session_id": "SNEVHB23H1KSZW4FJ15T", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.851+0000 7fe4ececbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146164854847, "job": 1, "event": "recovery_finished"} 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.851+0000 7fe4ececbd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.855+0000 7fe4ececbd80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.855+0000 7fe4ececbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55fc05114e00 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.855+0000 7fe4ececbd80 4 rocksdb: DB pointer 0x55fc05222000 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 0 starting mon.b rank 1 at public addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] at bind addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 1 mon.b@-1(???) e2 preinit fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 0 mon.b@-1(???).mds e1 new map 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 0 mon.b@-1(???).mds e1 print_map 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: e1 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: btime 2026-03-10T12:32:36:745933+0000 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: legacy client fscid: -1 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: No filesystems configured 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 0 mon.b@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:05.168 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:04 vm05 bash[29030]: debug 2026-03-10T12:36:04.859+0000 7fe4ececbd80 1 mon.b@-1(???).paxosservice(auth 1..10) refresh upgraded, format 0 -> 3 2026-03-10T12:36:05.383 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:36:05.383 INFO:teuthology.orchestra.run.vm01.stdout:[{"events": ["2026-03-10T12:33:19.615978Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-10T12:32:58.048040Z", "last_refresh": "2026-03-10T12:35:51.024563Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-10T12:36:01.853431Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-10T12:35:42.054947Z", "last_refresh": "2026-03-10T12:35:53.144992Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-10T12:35:43.517376Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-10T12:35:42.050980Z", "last_refresh": "2026-03-10T12:35:51.024486Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:33:46.420849Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm01=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T12:33:45.549327Z", "last_refresh": "2026-03-10T12:35:51.024543Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:35:42.069424Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T12:35:42.045863Z", "running": 0, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-10T12:35:51.024592Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: cluster 2026-03-10T12:34:16.475583+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: cluster 2026-03-10T12:34:16.475583+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: cephadm 2026-03-10T12:36:03.575178+0000 mgr.a (mgr.14150) 152 : cephadm [INF] Redeploying mon.b, (container cli args changed) . . . 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: cephadm 2026-03-10T12:36:03.575178+0000 mgr.a (mgr.14150) 152 : cephadm [INF] Redeploying mon.b, (container cli args changed) . . . 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: cephadm 2026-03-10T12:36:03.579192+0000 mgr.a (mgr.14150) 153 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: cephadm 2026-03-10T12:36:03.579192+0000 mgr.a (mgr.14150) 153 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: audit 2026-03-10T12:36:04.048845+0000 mon.a (mon.0) 18 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: audit 2026-03-10T12:36:04.048845+0000 mon.a (mon.0) 18 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: cluster 2026-03-10T12:36:04.054424+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: cluster 2026-03-10T12:36:04.054424+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: audit 2026-03-10T12:36:04.207163+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: audit 2026-03-10T12:36:04.207163+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: audit 2026-03-10T12:36:04.207257+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:05.400 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: audit 2026-03-10T12:36:04.207257+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:05.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: cluster 2026-03-10T12:34:16.475583+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:36:05.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: cluster 2026-03-10T12:34:16.475583+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: cephadm 2026-03-10T12:36:03.575178+0000 mgr.a (mgr.14150) 152 : cephadm [INF] Redeploying mon.b, (container cli args changed) . . . 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: cephadm 2026-03-10T12:36:03.575178+0000 mgr.a (mgr.14150) 152 : cephadm [INF] Redeploying mon.b, (container cli args changed) . . . 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: cephadm 2026-03-10T12:36:03.579192+0000 mgr.a (mgr.14150) 153 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: cephadm 2026-03-10T12:36:03.579192+0000 mgr.a (mgr.14150) 153 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: audit 2026-03-10T12:36:04.048845+0000 mon.a (mon.0) 18 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: audit 2026-03-10T12:36:04.048845+0000 mon.a (mon.0) 18 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: cluster 2026-03-10T12:36:04.054424+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: cluster 2026-03-10T12:36:04.054424+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: audit 2026-03-10T12:36:04.207163+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: audit 2026-03-10T12:36:04.207163+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: audit 2026-03-10T12:36:04.207257+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:05.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:05 vm05 bash[29030]: audit 2026-03-10T12:36:04.207257+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:05.473 INFO:tasks.cephadm:container.bar has 0/2 2026-03-10T12:36:06.067 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:06.067 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 systemd[1]: Stopping Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:36:06.067 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: debug 2026-03-10T12:36:05.852+0000 7f2ed439b640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-10T12:36:06.067 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34421]: debug 2026-03-10T12:36:05.852+0000 7f2ed439b640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-10T12:36:06.067 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 bash[34915]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-a 2026-03-10T12:36:06.067 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 systemd[1]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a.service: Deactivated successfully. 2026-03-10T12:36:06.067 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:05 vm01 systemd[1]: Stopped Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:36:06.067 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:36:05 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:06.067 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:36:05 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:06.449 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:36:06 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 systemd[1]: Started Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.240+0000 7f841ca78d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.240+0000 7f841ca78d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.240+0000 7f841ca78d80 0 pidfile_write: ignore empty --pid-file 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.240+0000 7f841ca78d80 0 load: jerasure load: lrc 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Git sha 0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: DB SUMMARY 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: DB Session ID: 3TPU4E4LLUF42BX60BHT 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: MANIFEST file: MANIFEST-000020 size: 373 Bytes 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 3, files: 000008.sst 000013.sst 000018.sst 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000019.log size: 144834 ; 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.env: 0x55f9e1b89dc0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.info_log: 0x55f9f0b98d00 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.statistics: (nil) 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.use_fsync: 0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T12:36:06.453 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.db_log_dir: 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.wal_dir: 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.write_buffer_manager: 0x55f9f0b9d900 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.unordered_write: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.row_cache: None 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.wal_filter: None 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.wal_compression: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_open_files: -1 2026-03-10T12:36:06.454 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Compression algorithms supported: 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: kZSTD supported: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.merge_operator: 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_filter: None 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55f9f0b98480) 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cache_index_and_filter_blocks: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: pin_top_level_index_and_filter: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: index_type: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: data_block_index_type: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: index_shortening: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: checksum: 4 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: no_block_cache: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: block_cache: 0x55f9f0bbf350 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: block_cache_name: BinnedLRUCache 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: block_cache_options: 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: capacity : 536870912 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: num_shard_bits : 4 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: strict_capacity_limit : 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: high_pri_pool_ratio: 0.000 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: block_cache_compressed: (nil) 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: persistent_cache: (nil) 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: block_size: 4096 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: block_size_deviation: 10 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: block_restart_interval: 16 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: index_block_restart_interval: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: metadata_block_size: 4096 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: partition_filters: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: use_delta_encoding: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: filter_policy: bloomfilter 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: whole_key_filtering: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: verify_compression: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: read_amp_bytes_per_bit: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: format_version: 5 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: enable_index_compression: 1 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: block_align: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: max_auto_readahead_size: 262144 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: prepopulate_block_cache: 0 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: initial_auto_readahead_size: 8192 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: num_file_reads_for_auto_readahead: 2 2026-03-10T12:36:06.455 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression: NoCompression 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.num_levels: 7 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T12:36:06.456 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.ttl: 2592000 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 succeeded,manifest_file_number is 20, next_file_number is 22, last_sequence is 3867, log_number is 15,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 15 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 15 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 38f355c1-e07f-4226-a178-26f487d5fa48 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146166249383, "job": 1, "event": "recovery_started", "wal_files": [19]} 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.244+0000 7f841ca78d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #19 mode 2 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.248+0000 7f841ca78d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146166251854, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 23, "file_size": 139307, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 3873, "largest_seqno": 3991, "table_properties": {"data_size": 137547, "index_size": 535, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 197, "raw_key_size": 1868, "raw_average_key_size": 26, "raw_value_size": 135769, "raw_average_value_size": 1912, "num_data_blocks": 20, "num_entries": 71, "num_filter_entries": 71, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773146166, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "38f355c1-e07f-4226-a178-26f487d5fa48", "db_session_id": "3TPU4E4LLUF42BX60BHT", "orig_file_number": 23, "seqno_to_time_mapping": "N/A"}} 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.248+0000 7f841ca78d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146166251918, "job": 1, "event": "recovery_finished"} 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.248+0000 7f841ca78d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 25 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000019.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55f9f0bc0e00 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 4 rocksdb: DB pointer 0x55f9f0cce000 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841a852640 4 rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 3] Compacting 4@0 files to L6, score 1.00 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841a852640 4 rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 2 Base level 0, inputs: [23(136KB) 18(3861KB) 13(70KB) 8(1900B)] 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841a852640 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146166257702, "job": 3, "event": "compaction_started", "compaction_reason": "LevelL0FilesNum", "files_L0": [23, 18, 13, 8], "score": 1, "input_data_size": 4168108, "oldest_snapshot_seqno": -1} 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 10 CompressorRegistry(0x55f9f0b87418) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f8419850640 2 Event(0x55f9f0c022c8 nevent=5000 time_id=1).set_owner center_id=2 owner=140205340558912 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f8419850640 10 stack operator() starting 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f8414045640 2 Event(0x55f9f0c02508 nevent=5000 time_id=1).set_owner center_id=1 owner=140205248239168 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f8414045640 10 stack operator() starting 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f8413844640 2 Event(0x55f9f0bb9dc8 nevent=5000 time_id=1).set_owner center_id=0 owner=140205239846464 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f8413844640 10 stack operator() starting 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] at bind addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 10 CompressorRegistry(0x55f9f0b87d18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 1 mon.a@-1(???) e2 preinit fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 0 mon.a@-1(???).mds e1 new map 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.252+0000 7f841ca78d80 0 mon.a@-1(???).mds e1 print_map 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: e1 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: btime 2026-03-10T12:32:36:745933+0000 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: legacy client fscid: -1 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: No filesystems configured 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.256+0000 7f841ca78d80 0 mon.a@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.256+0000 7f841ca78d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.256+0000 7f841ca78d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.256+0000 7f841ca78d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:06.457 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: debug 2026-03-10T12:36:06.256+0000 7f841ca78d80 1 mon.a@-1(???).paxosservice(auth 1..10) refresh upgraded, format 0 -> 3 2026-03-10T12:36:06.458 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 10 12:36:06 vm01 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:06.474 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch ls -f json 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cephadm 2026-03-10T12:36:05.203399+0000 mgr.a (mgr.14150) 156 : cephadm [INF] Redeploying mon.a, (entrypoint args changed) . . . 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cephadm 2026-03-10T12:36:05.203399+0000 mgr.a (mgr.14150) 156 : cephadm [INF] Redeploying mon.a, (entrypoint args changed) . . . 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cephadm 2026-03-10T12:36:05.211433+0000 mgr.a (mgr.14150) 157 : cephadm [INF] Deploying daemon mon.a on vm01 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cephadm 2026-03-10T12:36:05.211433+0000 mgr.a (mgr.14150) 157 : cephadm [INF] Deploying daemon mon.a on vm01 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:05.379814+0000 mgr.a (mgr.14150) 158 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:05.379814+0000 mgr.a (mgr.14150) 158 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:05.868677+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:05.868677+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:05.869149+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:05.869149+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.054685+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.054685+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.869575+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.869575+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.873262+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.873262+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878200+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878200+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878244+0000 mon.a (mon.0) 4 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878244+0000 mon.a (mon.0) 4 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878284+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878284+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878324+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:07.196 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878324+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878363+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878363+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878403+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878403+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878441+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878441+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878484+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878484+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878891+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878891+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878969+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.878969+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.880020+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.880020+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.880328+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cluster 2026-03-10T12:36:06.880328+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.889876+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.889876+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.894791+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.894791+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cephadm 2026-03-10T12:36:06.896205+0000 mgr.a (mgr.14150) 160 : cephadm [INF] Redeploying mon.b, (entrypoint args changed) . . . 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cephadm 2026-03-10T12:36:06.896205+0000 mgr.a (mgr.14150) 160 : cephadm [INF] Redeploying mon.b, (entrypoint args changed) . . . 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.896514+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.896514+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.897089+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.897089+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.897597+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: audit 2026-03-10T12:36:06.897597+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cephadm 2026-03-10T12:36:06.898112+0000 mgr.a (mgr.14150) 161 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:36:07.197 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:06 vm05 bash[29030]: cephadm 2026-03-10T12:36:06.898112+0000 mgr.a (mgr.14150) 161 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cephadm 2026-03-10T12:36:05.203399+0000 mgr.a (mgr.14150) 156 : cephadm [INF] Redeploying mon.a, (entrypoint args changed) . . . 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cephadm 2026-03-10T12:36:05.203399+0000 mgr.a (mgr.14150) 156 : cephadm [INF] Redeploying mon.a, (entrypoint args changed) . . . 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cephadm 2026-03-10T12:36:05.211433+0000 mgr.a (mgr.14150) 157 : cephadm [INF] Deploying daemon mon.a on vm01 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cephadm 2026-03-10T12:36:05.211433+0000 mgr.a (mgr.14150) 157 : cephadm [INF] Deploying daemon mon.a on vm01 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:05.379814+0000 mgr.a (mgr.14150) 158 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:05.379814+0000 mgr.a (mgr.14150) 158 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:05.868677+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:05.868677+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:05.869149+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:05.869149+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.054685+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.054685+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.869575+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.869575+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.873262+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.873262+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878200+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878200+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878244+0000 mon.a (mon.0) 4 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878244+0000 mon.a (mon.0) 4 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878284+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878284+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878324+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878324+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878363+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878363+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878403+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878403+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878441+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:07.199 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878441+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878484+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878484+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878891+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878891+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878969+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.878969+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.880020+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.880020+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.880328+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cluster 2026-03-10T12:36:06.880328+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.889876+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.889876+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.894791+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.894791+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cephadm 2026-03-10T12:36:06.896205+0000 mgr.a (mgr.14150) 160 : cephadm [INF] Redeploying mon.b, (entrypoint args changed) . . . 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cephadm 2026-03-10T12:36:06.896205+0000 mgr.a (mgr.14150) 160 : cephadm [INF] Redeploying mon.b, (entrypoint args changed) . . . 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.896514+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.896514+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.897089+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.897089+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.897597+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: audit 2026-03-10T12:36:06.897597+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cephadm 2026-03-10T12:36:06.898112+0000 mgr.a (mgr.14150) 161 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:36:07.200 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:06 vm01 bash[35029]: cephadm 2026-03-10T12:36:06.898112+0000 mgr.a (mgr.14150) 161 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-10T12:36:07.508 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:07 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:07.508 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:07 vm05 systemd[1]: Stopping Ceph mon.b for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:36:07.508 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:36:07 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:07.508 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:36:07 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:08.121 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:07 vm05 bash[29030]: debug 2026-03-10T12:36:07.843+0000 7fe4eb4a6640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-10T12:36:08.122 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:07 vm05 bash[29030]: debug 2026-03-10T12:36:07.843+0000 7fe4eb4a6640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-10T12:36:08.122 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:07 vm05 bash[29299]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-b 2026-03-10T12:36:08.122 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:07 vm05 bash[29354]: Error response from daemon: No such container: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-b 2026-03-10T12:36:08.122 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:07 vm05 systemd[1]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.b.service: Deactivated successfully. 2026-03-10T12:36:08.122 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:07 vm05 systemd[1]: Stopped Ceph mon.b for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:36:08.424 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 10 12:36:08 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 systemd[1]: Started Ceph mon.b for f864ef90-1c7c-11f1-946a-0b343fff599a. 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.323+0000 7f1e4c4fbd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.323+0000 7f1e4c4fbd80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.323+0000 7f1e4c4fbd80 0 pidfile_write: ignore empty --pid-file 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.323+0000 7f1e4c4fbd80 0 load: jerasure load: lrc 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Git sha 0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: DB SUMMARY 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: DB Session ID: W21XIGJ1GMAVXABCKTPY 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: MANIFEST file: MANIFEST-000015 size: 282 Bytes 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000014.log size: 191194 ; 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.env: 0x561f861bddc0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.info_log: 0x561fb700cd00 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.statistics: (nil) 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.use_fsync: 0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T12:36:08.427 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.db_log_dir: 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.wal_dir: 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.write_buffer_manager: 0x561fb7011900 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.unordered_write: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.row_cache: None 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.wal_filter: None 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.wal_compression: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T12:36:08.428 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_open_files: -1 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Compression algorithms supported: 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: kZSTD supported: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.merge_operator: 2026-03-10T12:36:08.429 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_filter: None 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x561fb700c480) 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: cache_index_and_filter_blocks: 1 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: pin_top_level_index_and_filter: 1 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: index_type: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: data_block_index_type: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: index_shortening: 1 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: checksum: 4 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: no_block_cache: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: block_cache: 0x561fb7033350 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: block_cache_name: BinnedLRUCache 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: block_cache_options: 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: capacity : 536870912 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: num_shard_bits : 4 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: strict_capacity_limit : 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: high_pri_pool_ratio: 0.000 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: block_cache_compressed: (nil) 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: persistent_cache: (nil) 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: block_size: 4096 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: block_size_deviation: 10 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: block_restart_interval: 16 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: index_block_restart_interval: 1 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: metadata_block_size: 4096 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: partition_filters: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: use_delta_encoding: 1 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: filter_policy: bloomfilter 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: whole_key_filtering: 1 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: verify_compression: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: read_amp_bytes_per_bit: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: format_version: 5 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: enable_index_compression: 1 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: block_align: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: max_auto_readahead_size: 262144 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: prepopulate_block_cache: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: initial_auto_readahead_size: 8192 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: num_file_reads_for_auto_readahead: 2 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression: NoCompression 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.num_levels: 7 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T12:36:08.430 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.ttl: 2592000 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T12:36:08.431 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.327+0000 7f1e4c4fbd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.335+0000 7f1e4c4fbd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 3831, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.335+0000 7f1e4c4fbd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.335+0000 7f1e4c4fbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 6bd5b8b6-8815-4e14-b03d-5213c86c0db1 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.335+0000 7f1e4c4fbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146168335869, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.335+0000 7f1e4c4fbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.335+0000 7f1e4c4fbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146168337803, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 125541, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 3843, "largest_seqno": 3950, "table_properties": {"data_size": 123856, "index_size": 460, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 197, "raw_key_size": 1865, "raw_average_key_size": 26, "raw_value_size": 122164, "raw_average_value_size": 1745, "num_data_blocks": 18, "num_entries": 70, "num_filter_entries": 70, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773146168, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "6bd5b8b6-8815-4e14-b03d-5213c86c0db1", "db_session_id": "W21XIGJ1GMAVXABCKTPY", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.335+0000 7f1e4c4fbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773146168337892, "job": 1, "event": "recovery_finished"} 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.335+0000 7f1e4c4fbd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.339+0000 7f1e4c4fbd80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.339+0000 7f1e4c4fbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x561fb7034e00 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.339+0000 7f1e4c4fbd80 4 rocksdb: DB pointer 0x561fb7142000 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.339+0000 7f1e422c5640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.339+0000 7f1e422c5640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: ** DB Stats ** 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: ** Compaction Stats [default] ** 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: L0 3/0 4.01 MB 0.8 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 71.1 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Sum 3/0 4.01 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 71.1 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 71.1 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: ** Compaction Stats [default] ** 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 71.1 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: AddFile(Keys): cumulative 0, interval 0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Cumulative compaction: 0.00 GB write, 9.01 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Interval compaction: 0.00 GB write, 9.01 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Block cache BinnedLRUCache@0x561fb7033350#7 capacity: 512.00 MB usage: 14.61 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 9e-06 secs_since: 0 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: Block cache entry stats(count,size,portion): FilterBlock(3,4.88 KB,0.000929832%) IndexBlock(3,9.73 KB,0.00185668%) Misc(1,0.00 KB,0%) 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: ** File Read Latency Histogram By Level [default] ** 2026-03-10T12:36:08.432 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:36:08 vm05 systemd[1]: /etc/systemd/system/ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T12:36:08.432 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.339+0000 7f1e4c4fbd80 10 CompressorRegistry(0x561fb6ffb418) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.339+0000 7f1e432c7640 2 Event(0x561fb702ddc8 nevent=5000 time_id=1).set_owner center_id=0 owner=139767952733760 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.339+0000 7f1e43ac8640 2 Event(0x561fb7062088 nevent=5000 time_id=1).set_owner center_id=1 owner=139767961126464 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e432c7640 10 stack operator() starting 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e43ac8640 10 stack operator() starting 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e492d3640 2 Event(0x561fb70622c8 nevent=5000 time_id=1).set_owner center_id=2 owner=139768053446208 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e492d3640 10 stack operator() starting 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 0 starting mon.b rank 1 at public addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] at bind addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 10 CompressorRegistry(0x561fb6ffbd18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 1 mon.b@-1(???) e2 preinit fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 0 mon.b@-1(???).mds e1 new map 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 0 mon.b@-1(???).mds e1 print_map 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: e1 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: btime 2026-03-10T12:32:36:745933+0000 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: legacy client fscid: -1 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: No filesystems configured 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 0 mon.b@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:36:08.433 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:08 vm05 bash[29416]: debug 2026-03-10T12:36:08.343+0000 7f1e4c4fbd80 1 mon.b@-1(???).paxosservice(auth 1..11) refresh upgraded, format 0 -> 3 2026-03-10T12:36:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.054951+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.054951+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.478381+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T12:36:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.478381+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T12:36:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.481470+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.481470+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.484338+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.484338+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489190+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489190+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489201+0000 mon.a (mon.0) 22 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489201+0000 mon.a (mon.0) 22 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489206+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489206+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489220+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489220+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489225+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489225+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489230+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489230+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489346+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489346+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489350+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489350+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489737+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489737+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489756+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.489756+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.490016+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.490016+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.490105+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cluster 2026-03-10T12:36:08.490105+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.494092+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.494092+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.498453+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.498453+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.499724+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.499724+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.500831+0000 mon.a (mon.0) 36 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.500831+0000 mon.a (mon.0) 36 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.501258+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.501258+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.505358+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: audit 2026-03-10T12:36:08.505358+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cephadm 2026-03-10T12:36:08.505719+0000 mgr.a (mgr.14150) 163 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:36:09.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:09 vm05 bash[29416]: cephadm 2026-03-10T12:36:08.505719+0000 mgr.a (mgr.14150) 163 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.054951+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.054951+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.478381+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.478381+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.481470+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.481470+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.484338+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.484338+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489190+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489190+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489201+0000 mon.a (mon.0) 22 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489201+0000 mon.a (mon.0) 22 : cluster [DBG] fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489206+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489206+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-10T12:33:30.138463+0000 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489220+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489220+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-10T12:32:35.364821+0000 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489225+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489225+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489230+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489230+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489346+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489346+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489350+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489350+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489737+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489737+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489756+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.489756+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.490016+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.490016+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.490105+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cluster 2026-03-10T12:36:08.490105+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-10T12:36:09.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.494092+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.494092+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.498453+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.498453+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.499724+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.499724+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.500831+0000 mon.a (mon.0) 36 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.500831+0000 mon.a (mon.0) 36 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.501258+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.501258+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.505358+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: audit 2026-03-10T12:36:08.505358+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cephadm 2026-03-10T12:36:08.505719+0000 mgr.a (mgr.14150) 163 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:36:09.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:09 vm01 bash[35029]: cephadm 2026-03-10T12:36:08.505719+0000 mgr.a (mgr.14150) 163 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T12:36:10.109 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:36:10.428 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:36:10.428 INFO:teuthology.orchestra.run.vm01.stdout:[{"events": ["2026-03-10T12:33:19.615978Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-10T12:32:58.048040Z", "last_refresh": "2026-03-10T12:35:51.024563Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-10T12:36:01.853431Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-10T12:35:42.054947Z", "last_refresh": "2026-03-10T12:35:53.144992Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-10T12:35:43.517376Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-10T12:35:42.050980Z", "last_refresh": "2026-03-10T12:35:51.024486Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:33:46.420849Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm01=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T12:33:45.549327Z", "last_refresh": "2026-03-10T12:35:51.024543Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:35:42.069424Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T12:35:42.045863Z", "running": 0, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-10T12:35:51.024592Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-10T12:36:10.491 INFO:tasks.cephadm:container.bar has 0/2 2026-03-10T12:36:11.492 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph orch ls -f json 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: cluster 2026-03-10T12:36:10.055233+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: cluster 2026-03-10T12:36:10.055233+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.424223+0000 mgr.a (mgr.14150) 165 : audit [DBG] from='client.44100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.424223+0000 mgr.a (mgr.14150) 165 : audit [DBG] from='client.44100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.494104+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.494104+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.499459+0000 mon.a (mon.0) 40 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.499459+0000 mon.a (mon.0) 40 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.503260+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.503260+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.637986+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.637986+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.642203+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.642203+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.643474+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.643474+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.645213+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.645213+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.646513+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.646513+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.647125+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.647125+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.650304+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.925 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:11 vm05 bash[29416]: audit 2026-03-10T12:36:10.650304+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: cluster 2026-03-10T12:36:10.055233+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: cluster 2026-03-10T12:36:10.055233+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.424223+0000 mgr.a (mgr.14150) 165 : audit [DBG] from='client.44100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.424223+0000 mgr.a (mgr.14150) 165 : audit [DBG] from='client.44100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.494104+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.494104+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.499459+0000 mon.a (mon.0) 40 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.499459+0000 mon.a (mon.0) 40 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.503260+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.503260+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.637986+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.637986+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.642203+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.642203+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.643474+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.643474+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.645213+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.645213+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.646513+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.646513+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.647125+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.647125+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.650304+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:11.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:11 vm01 bash[35029]: audit 2026-03-10T12:36:10.650304+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:13.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:13 vm05 bash[29416]: cluster 2026-03-10T12:36:12.055540+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:13.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:13 vm05 bash[29416]: cluster 2026-03-10T12:36:12.055540+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:13.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:13 vm01 bash[35029]: cluster 2026-03-10T12:36:12.055540+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:13.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:13 vm01 bash[35029]: cluster 2026-03-10T12:36:12.055540+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:15.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:15 vm05 bash[29416]: cluster 2026-03-10T12:36:14.055824+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:15.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:15 vm05 bash[29416]: cluster 2026-03-10T12:36:14.055824+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:15.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:15 vm01 bash[35029]: cluster 2026-03-10T12:36:14.055824+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:15.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:15 vm01 bash[35029]: cluster 2026-03-10T12:36:14.055824+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:16.117 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:36:16.404 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T12:36:16.404 INFO:teuthology.orchestra.run.vm01.stdout:[{"events": ["2026-03-10T12:33:19.615978Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-10T12:32:58.048040Z", "last_refresh": "2026-03-10T12:36:10.487713Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-10T12:36:01.853431Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-10T12:35:42.054947Z", "last_refresh": "2026-03-10T12:36:10.487808Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-10T12:35:43.517376Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-10T12:35:42.050980Z", "last_refresh": "2026-03-10T12:36:10.487631Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:33:46.420849Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm01=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T12:33:45.549327Z", "last_refresh": "2026-03-10T12:36:10.487692Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T12:35:42.069424Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T12:35:42.045863Z", "last_refresh": "2026-03-10T12:36:10.487772Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-10T12:36:10.487742Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-10T12:36:16.461 INFO:tasks.cephadm:container.bar has 2/2 2026-03-10T12:36:16.461 INFO:teuthology.run_tasks:Running task exec... 2026-03-10T12:36:16.464 INFO:teuthology.task.exec:Executing custom commands... 2026-03-10T12:36:16.464 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'set -ex 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> sleep 60 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # check extra container and entrypoint args written to mon unit run file 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # check that custom container properly wrote content to file. 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # This requires the custom config, extra container args, and 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # entrypoint args to all be working in order for this to have 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # been written. The container entrypoint was set up with custom_configs, 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # the content and where to write to with the entrypoint args, and the mounting 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # of the /etc/cephadm_testing dir with extra container args 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> grep "testing_custom_containers" /etc/cephadm_testing/testing.txt 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # Verify that container bar'"'"'s init containers and primary container 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> # ran successfully 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name '"'"'container.bar.*'"'"') 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> test -n "$dir" 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> grep ok ${dir}/data/primary.txt 2026-03-10T12:36:16.464 DEBUG:teuthology.orchestra.run.vm01:> grep from=centos ${dir}/data/from.txt 2026-03-10T12:36:16.465 DEBUG:teuthology.orchestra.run.vm01:> test -s ${dir}/data/presized.dat 2026-03-10T12:36:16.465 DEBUG:teuthology.orchestra.run.vm01:> ' 2026-03-10T12:36:16.473 INFO:teuthology.orchestra.run.vm01.stderr:++ /home/ubuntu/cephtest/cephadm shell -- ceph fsid 2026-03-10T12:36:17.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:17 vm05 bash[29416]: cluster 2026-03-10T12:36:16.056118+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:17.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:17 vm05 bash[29416]: cluster 2026-03-10T12:36:16.056118+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:17.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:17 vm05 bash[29416]: audit 2026-03-10T12:36:16.402412+0000 mgr.a (mgr.14150) 169 : audit [DBG] from='client.44104 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:17.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:17 vm05 bash[29416]: audit 2026-03-10T12:36:16.402412+0000 mgr.a (mgr.14150) 169 : audit [DBG] from='client.44104 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:17.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:17 vm01 bash[35029]: cluster 2026-03-10T12:36:16.056118+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:17.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:17 vm01 bash[35029]: cluster 2026-03-10T12:36:16.056118+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:17.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:17 vm01 bash[35029]: audit 2026-03-10T12:36:16.402412+0000 mgr.a (mgr.14150) 169 : audit [DBG] from='client.44104 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:17.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:17 vm01 bash[35029]: audit 2026-03-10T12:36:16.402412+0000 mgr.a (mgr.14150) 169 : audit [DBG] from='client.44104 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:36:19.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:19 vm05 bash[29416]: cluster 2026-03-10T12:36:18.056433+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:19.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:19 vm05 bash[29416]: cluster 2026-03-10T12:36:18.056433+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:19.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:19 vm01 bash[35029]: cluster 2026-03-10T12:36:18.056433+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:19.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:19 vm01 bash[35029]: cluster 2026-03-10T12:36:18.056433+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:20.135 INFO:teuthology.orchestra.run.vm01.stderr:Inferring fsid f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:21.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:21 vm05 bash[29416]: cluster 2026-03-10T12:36:20.056697+0000 mgr.a (mgr.14150) 171 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:21.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:21 vm05 bash[29416]: cluster 2026-03-10T12:36:20.056697+0000 mgr.a (mgr.14150) 171 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:21.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:21 vm01 bash[35029]: cluster 2026-03-10T12:36:20.056697+0000 mgr.a (mgr.14150) 171 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:21.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:21 vm01 bash[35029]: cluster 2026-03-10T12:36:20.056697+0000 mgr.a (mgr.14150) 171 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:24.150 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:23 vm01 bash[35029]: cluster 2026-03-10T12:36:22.056921+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:24.150 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:23 vm01 bash[35029]: cluster 2026-03-10T12:36:22.056921+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:24.153 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:36:24.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:23 vm05 bash[29416]: cluster 2026-03-10T12:36:22.056921+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:24.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:23 vm05 bash[29416]: cluster 2026-03-10T12:36:22.056921+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:26.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:25 vm05 bash[29416]: cluster 2026-03-10T12:36:24.057153+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:26.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:25 vm05 bash[29416]: cluster 2026-03-10T12:36:24.057153+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:26.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:25 vm01 bash[35029]: cluster 2026-03-10T12:36:24.057153+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:26.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:25 vm01 bash[35029]: cluster 2026-03-10T12:36:24.057153+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: cluster 2026-03-10T12:36:26.057424+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: cluster 2026-03-10T12:36:26.057424+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.206482+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.206482+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.212766+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.212766+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.217627+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.217627+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.298095+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.298095+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.302043+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.302043+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.306020+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:27 vm05 bash[29416]: audit 2026-03-10T12:36:27.306020+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.182 INFO:teuthology.orchestra.run.vm01.stderr:Using ceph image with id '654f31e6858e' and tag 'e911bdebe5c8faa3800735d1568fcdca65db60df' created on 2026-02-25 18:57:17 +0000 UTC 2026-03-10T12:36:28.182 INFO:teuthology.orchestra.run.vm01.stderr:quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: cluster 2026-03-10T12:36:26.057424+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: cluster 2026-03-10T12:36:26.057424+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.206482+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.206482+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.212766+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.212766+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.217627+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.217627+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.298095+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.298095+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.302043+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.302043+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.306020+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.195 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:27 vm01 bash[35029]: audit 2026-03-10T12:36:27.306020+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:28.504 INFO:teuthology.orchestra.run.vm01.stderr:+ FSID=f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:36:28.504 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 60 2026-03-10T12:36:28.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:28 vm01 bash[35029]: audit 2026-03-10T12:36:28.455712+0000 mon.a (mon.0) 55 : audit [DBG] from='client.? 192.168.123.101:0/497162004' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-10T12:36:28.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:28 vm01 bash[35029]: audit 2026-03-10T12:36:28.455712+0000 mon.a (mon.0) 55 : audit [DBG] from='client.? 192.168.123.101:0/497162004' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-10T12:36:29.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:28 vm05 bash[29416]: audit 2026-03-10T12:36:28.455712+0000 mon.a (mon.0) 55 : audit [DBG] from='client.? 192.168.123.101:0/497162004' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-10T12:36:29.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:28 vm05 bash[29416]: audit 2026-03-10T12:36:28.455712+0000 mon.a (mon.0) 55 : audit [DBG] from='client.? 192.168.123.101:0/497162004' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-10T12:36:30.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:29 vm05 bash[29416]: cluster 2026-03-10T12:36:28.057685+0000 mgr.a (mgr.14150) 175 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:30.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:29 vm05 bash[29416]: cluster 2026-03-10T12:36:28.057685+0000 mgr.a (mgr.14150) 175 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:30.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:29 vm01 bash[35029]: cluster 2026-03-10T12:36:28.057685+0000 mgr.a (mgr.14150) 175 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:30.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:29 vm01 bash[35029]: cluster 2026-03-10T12:36:28.057685+0000 mgr.a (mgr.14150) 175 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:32.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:31 vm05 bash[29416]: cluster 2026-03-10T12:36:30.057956+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:32.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:31 vm05 bash[29416]: cluster 2026-03-10T12:36:30.057956+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:32.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:31 vm01 bash[35029]: cluster 2026-03-10T12:36:30.057956+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:32.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:31 vm01 bash[35029]: cluster 2026-03-10T12:36:30.057956+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:34.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:33 vm05 bash[29416]: cluster 2026-03-10T12:36:32.058222+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:34.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:33 vm05 bash[29416]: cluster 2026-03-10T12:36:32.058222+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:34.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:33 vm01 bash[35029]: cluster 2026-03-10T12:36:32.058222+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:34.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:33 vm01 bash[35029]: cluster 2026-03-10T12:36:32.058222+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:36.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:35 vm05 bash[29416]: cluster 2026-03-10T12:36:34.058491+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:36.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:35 vm05 bash[29416]: cluster 2026-03-10T12:36:34.058491+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:36.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:35 vm01 bash[35029]: cluster 2026-03-10T12:36:34.058491+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:36.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:35 vm01 bash[35029]: cluster 2026-03-10T12:36:34.058491+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:38.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:37 vm01 bash[35029]: cluster 2026-03-10T12:36:36.058684+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:38.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:37 vm01 bash[35029]: cluster 2026-03-10T12:36:36.058684+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:38.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:37 vm05 bash[29416]: cluster 2026-03-10T12:36:36.058684+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:38.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:38 vm05 bash[29416]: cluster 2026-03-10T12:36:36.058684+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:39.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:39 vm05 bash[29416]: cluster 2026-03-10T12:36:38.058870+0000 mgr.a (mgr.14150) 180 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:39.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:39 vm05 bash[29416]: cluster 2026-03-10T12:36:38.058870+0000 mgr.a (mgr.14150) 180 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:39.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:39 vm01 bash[35029]: cluster 2026-03-10T12:36:38.058870+0000 mgr.a (mgr.14150) 180 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:39.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:39 vm01 bash[35029]: cluster 2026-03-10T12:36:38.058870+0000 mgr.a (mgr.14150) 180 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:41.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:41 vm05 bash[29416]: cluster 2026-03-10T12:36:40.059170+0000 mgr.a (mgr.14150) 181 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:41.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:41 vm05 bash[29416]: cluster 2026-03-10T12:36:40.059170+0000 mgr.a (mgr.14150) 181 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:41.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:41 vm01 bash[35029]: cluster 2026-03-10T12:36:40.059170+0000 mgr.a (mgr.14150) 181 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:41.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:41 vm01 bash[35029]: cluster 2026-03-10T12:36:40.059170+0000 mgr.a (mgr.14150) 181 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:43.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:43 vm05 bash[29416]: cluster 2026-03-10T12:36:42.059496+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:43.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:43 vm05 bash[29416]: cluster 2026-03-10T12:36:42.059496+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:43.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:43 vm01 bash[35029]: cluster 2026-03-10T12:36:42.059496+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:43.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:43 vm01 bash[35029]: cluster 2026-03-10T12:36:42.059496+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:45.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:45 vm05 bash[29416]: cluster 2026-03-10T12:36:44.059764+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:45.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:45 vm05 bash[29416]: cluster 2026-03-10T12:36:44.059764+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:45.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:45 vm01 bash[35029]: cluster 2026-03-10T12:36:44.059764+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:45.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:45 vm01 bash[35029]: cluster 2026-03-10T12:36:44.059764+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:47.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:47 vm05 bash[29416]: cluster 2026-03-10T12:36:46.060035+0000 mgr.a (mgr.14150) 184 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:47.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:47 vm05 bash[29416]: cluster 2026-03-10T12:36:46.060035+0000 mgr.a (mgr.14150) 184 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:47.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:47 vm05 bash[29416]: audit 2026-03-10T12:36:47.284093+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:47 vm05 bash[29416]: audit 2026-03-10T12:36:47.284093+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:47 vm05 bash[29416]: audit 2026-03-10T12:36:47.288280+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:47 vm05 bash[29416]: audit 2026-03-10T12:36:47.288280+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:47 vm05 bash[29416]: audit 2026-03-10T12:36:47.291820+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:47 vm05 bash[29416]: audit 2026-03-10T12:36:47.291820+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:47 vm01 bash[35029]: cluster 2026-03-10T12:36:46.060035+0000 mgr.a (mgr.14150) 184 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:47.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:47 vm01 bash[35029]: cluster 2026-03-10T12:36:46.060035+0000 mgr.a (mgr.14150) 184 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:47.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:47 vm01 bash[35029]: audit 2026-03-10T12:36:47.284093+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:47 vm01 bash[35029]: audit 2026-03-10T12:36:47.284093+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:47 vm01 bash[35029]: audit 2026-03-10T12:36:47.288280+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:47 vm01 bash[35029]: audit 2026-03-10T12:36:47.288280+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:47 vm01 bash[35029]: audit 2026-03-10T12:36:47.291820+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:47.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:47 vm01 bash[35029]: audit 2026-03-10T12:36:47.291820+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:48 vm05 bash[29416]: audit 2026-03-10T12:36:47.376281+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:48 vm05 bash[29416]: audit 2026-03-10T12:36:47.376281+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:48 vm05 bash[29416]: audit 2026-03-10T12:36:47.380008+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:48 vm05 bash[29416]: audit 2026-03-10T12:36:47.380008+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:48 vm05 bash[29416]: audit 2026-03-10T12:36:47.383480+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:48 vm05 bash[29416]: audit 2026-03-10T12:36:47.383480+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:48 vm01 bash[35029]: audit 2026-03-10T12:36:47.376281+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:48 vm01 bash[35029]: audit 2026-03-10T12:36:47.376281+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:48 vm01 bash[35029]: audit 2026-03-10T12:36:47.380008+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:48 vm01 bash[35029]: audit 2026-03-10T12:36:47.380008+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:48 vm01 bash[35029]: audit 2026-03-10T12:36:47.383480+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:48.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:48 vm01 bash[35029]: audit 2026-03-10T12:36:47.383480+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:36:49.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:49 vm05 bash[29416]: cluster 2026-03-10T12:36:48.060285+0000 mgr.a (mgr.14150) 185 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:49.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:49 vm05 bash[29416]: cluster 2026-03-10T12:36:48.060285+0000 mgr.a (mgr.14150) 185 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:49 vm01 bash[35029]: cluster 2026-03-10T12:36:48.060285+0000 mgr.a (mgr.14150) 185 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:49 vm01 bash[35029]: cluster 2026-03-10T12:36:48.060285+0000 mgr.a (mgr.14150) 185 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:51.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:51 vm05 bash[29416]: cluster 2026-03-10T12:36:50.060516+0000 mgr.a (mgr.14150) 186 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:51.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:51 vm05 bash[29416]: cluster 2026-03-10T12:36:50.060516+0000 mgr.a (mgr.14150) 186 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:51.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:51 vm01 bash[35029]: cluster 2026-03-10T12:36:50.060516+0000 mgr.a (mgr.14150) 186 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:51.987 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:51 vm01 bash[35029]: cluster 2026-03-10T12:36:50.060516+0000 mgr.a (mgr.14150) 186 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:53.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:53 vm05 bash[29416]: cluster 2026-03-10T12:36:52.060752+0000 mgr.a (mgr.14150) 187 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:53.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:53 vm05 bash[29416]: cluster 2026-03-10T12:36:52.060752+0000 mgr.a (mgr.14150) 187 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:53.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:53 vm01 bash[35029]: cluster 2026-03-10T12:36:52.060752+0000 mgr.a (mgr.14150) 187 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:53.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:53 vm01 bash[35029]: cluster 2026-03-10T12:36:52.060752+0000 mgr.a (mgr.14150) 187 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:55.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:55 vm05 bash[29416]: cluster 2026-03-10T12:36:54.060966+0000 mgr.a (mgr.14150) 188 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:55.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:55 vm05 bash[29416]: cluster 2026-03-10T12:36:54.060966+0000 mgr.a (mgr.14150) 188 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:55.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:55 vm01 bash[35029]: cluster 2026-03-10T12:36:54.060966+0000 mgr.a (mgr.14150) 188 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:55.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:55 vm01 bash[35029]: cluster 2026-03-10T12:36:54.060966+0000 mgr.a (mgr.14150) 188 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:57.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:57 vm05 bash[29416]: cluster 2026-03-10T12:36:56.061198+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:57.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:57 vm05 bash[29416]: cluster 2026-03-10T12:36:56.061198+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:57.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:57 vm01 bash[35029]: cluster 2026-03-10T12:36:56.061198+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:36:57.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:57 vm01 bash[35029]: cluster 2026-03-10T12:36:56.061198+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:00.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:59 vm05 bash[29416]: cluster 2026-03-10T12:36:58.061376+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:00.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:36:59 vm05 bash[29416]: cluster 2026-03-10T12:36:58.061376+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:00.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:59 vm01 bash[35029]: cluster 2026-03-10T12:36:58.061376+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:00.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:36:59 vm01 bash[35029]: cluster 2026-03-10T12:36:58.061376+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:01.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:00 vm01 bash[35029]: cluster 2026-03-10T12:37:00.061602+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:01.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:00 vm01 bash[35029]: cluster 2026-03-10T12:37:00.061602+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:01.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:00 vm05 bash[29416]: cluster 2026-03-10T12:37:00.061602+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:01.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:00 vm05 bash[29416]: cluster 2026-03-10T12:37:00.061602+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:03 vm05 bash[29416]: cluster 2026-03-10T12:37:02.061907+0000 mgr.a (mgr.14150) 192 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:03.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:03 vm05 bash[29416]: cluster 2026-03-10T12:37:02.061907+0000 mgr.a (mgr.14150) 192 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:03.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:03 vm01 bash[35029]: cluster 2026-03-10T12:37:02.061907+0000 mgr.a (mgr.14150) 192 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:03.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:03 vm01 bash[35029]: cluster 2026-03-10T12:37:02.061907+0000 mgr.a (mgr.14150) 192 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:05.570 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:05 vm01 bash[35029]: cluster 2026-03-10T12:37:04.062147+0000 mgr.a (mgr.14150) 193 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:05.570 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:05 vm01 bash[35029]: cluster 2026-03-10T12:37:04.062147+0000 mgr.a (mgr.14150) 193 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:05.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:05 vm05 bash[29416]: cluster 2026-03-10T12:37:04.062147+0000 mgr.a (mgr.14150) 193 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:05.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:05 vm05 bash[29416]: cluster 2026-03-10T12:37:04.062147+0000 mgr.a (mgr.14150) 193 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:07.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:07 vm05 bash[29416]: cluster 2026-03-10T12:37:06.062407+0000 mgr.a (mgr.14150) 194 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:07.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:07 vm05 bash[29416]: cluster 2026-03-10T12:37:06.062407+0000 mgr.a (mgr.14150) 194 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:07.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:07 vm01 bash[35029]: cluster 2026-03-10T12:37:06.062407+0000 mgr.a (mgr.14150) 194 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:07.711 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:07 vm01 bash[35029]: cluster 2026-03-10T12:37:06.062407+0000 mgr.a (mgr.14150) 194 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.366277+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.366277+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.370905+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.370905+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.376563+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.376563+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.458070+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.458070+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.462513+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.462513+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.468837+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.674 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:08 vm05 bash[29416]: audit 2026-03-10T12:37:07.468837+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.366277+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.366277+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.370905+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.370905+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.376563+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.376563+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.458070+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.458070+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.462513+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.462513+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.468837+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:08.698 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:08 vm01 bash[35029]: audit 2026-03-10T12:37:07.468837+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:09 vm05 bash[29416]: cluster 2026-03-10T12:37:08.062703+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:09.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:09 vm05 bash[29416]: cluster 2026-03-10T12:37:08.062703+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:09.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:09 vm01 bash[35029]: cluster 2026-03-10T12:37:08.062703+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:09.961 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:09 vm01 bash[35029]: cluster 2026-03-10T12:37:08.062703+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: cluster 2026-03-10T12:37:10.063012+0000 mgr.a (mgr.14150) 196 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: cluster 2026-03-10T12:37:10.063012+0000 mgr.a (mgr.14150) 196 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: audit 2026-03-10T12:37:10.661417+0000 mon.a (mon.0) 68 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: audit 2026-03-10T12:37:10.661417+0000 mon.a (mon.0) 68 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: audit 2026-03-10T12:37:10.662756+0000 mon.a (mon.0) 69 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: audit 2026-03-10T12:37:10.662756+0000 mon.a (mon.0) 69 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: audit 2026-03-10T12:37:10.663210+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: audit 2026-03-10T12:37:10.663210+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: audit 2026-03-10T12:37:10.668318+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:11.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:11 vm05 bash[29416]: audit 2026-03-10T12:37:10.668318+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: cluster 2026-03-10T12:37:10.063012+0000 mgr.a (mgr.14150) 196 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: cluster 2026-03-10T12:37:10.063012+0000 mgr.a (mgr.14150) 196 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: audit 2026-03-10T12:37:10.661417+0000 mon.a (mon.0) 68 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: audit 2026-03-10T12:37:10.661417+0000 mon.a (mon.0) 68 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: audit 2026-03-10T12:37:10.662756+0000 mon.a (mon.0) 69 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: audit 2026-03-10T12:37:10.662756+0000 mon.a (mon.0) 69 : audit [DBG] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: audit 2026-03-10T12:37:10.663210+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: audit 2026-03-10T12:37:10.663210+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: audit 2026-03-10T12:37:10.668318+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:11.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:11 vm01 bash[35029]: audit 2026-03-10T12:37:10.668318+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:13.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:13 vm05 bash[29416]: cluster 2026-03-10T12:37:12.063272+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:13.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:13 vm05 bash[29416]: cluster 2026-03-10T12:37:12.063272+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:13.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:13 vm01 bash[35029]: cluster 2026-03-10T12:37:12.063272+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:13.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:13 vm01 bash[35029]: cluster 2026-03-10T12:37:12.063272+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:15 vm05 bash[29416]: cluster 2026-03-10T12:37:14.063506+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:16.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:15 vm05 bash[29416]: cluster 2026-03-10T12:37:14.063506+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:16.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:15 vm01 bash[35029]: cluster 2026-03-10T12:37:14.063506+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:16.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:15 vm01 bash[35029]: cluster 2026-03-10T12:37:14.063506+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:17 vm05 bash[29416]: cluster 2026-03-10T12:37:16.063714+0000 mgr.a (mgr.14150) 199 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:18.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:17 vm05 bash[29416]: cluster 2026-03-10T12:37:16.063714+0000 mgr.a (mgr.14150) 199 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:18.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:17 vm01 bash[35029]: cluster 2026-03-10T12:37:16.063714+0000 mgr.a (mgr.14150) 199 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:18.210 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:17 vm01 bash[35029]: cluster 2026-03-10T12:37:16.063714+0000 mgr.a (mgr.14150) 199 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:20.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:19 vm05 bash[29416]: cluster 2026-03-10T12:37:18.063918+0000 mgr.a (mgr.14150) 200 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:20.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:19 vm05 bash[29416]: cluster 2026-03-10T12:37:18.063918+0000 mgr.a (mgr.14150) 200 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:20.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:19 vm01 bash[35029]: cluster 2026-03-10T12:37:18.063918+0000 mgr.a (mgr.14150) 200 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:20.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:19 vm01 bash[35029]: cluster 2026-03-10T12:37:18.063918+0000 mgr.a (mgr.14150) 200 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:22.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:21 vm05 bash[29416]: cluster 2026-03-10T12:37:20.064175+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:22.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:21 vm05 bash[29416]: cluster 2026-03-10T12:37:20.064175+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:22.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:21 vm01 bash[35029]: cluster 2026-03-10T12:37:20.064175+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:22.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:21 vm01 bash[35029]: cluster 2026-03-10T12:37:20.064175+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:24.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:23 vm01 bash[35029]: cluster 2026-03-10T12:37:22.064465+0000 mgr.a (mgr.14150) 202 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:24.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:23 vm01 bash[35029]: cluster 2026-03-10T12:37:22.064465+0000 mgr.a (mgr.14150) 202 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:24.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:23 vm05 bash[29416]: cluster 2026-03-10T12:37:22.064465+0000 mgr.a (mgr.14150) 202 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:24.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:23 vm05 bash[29416]: cluster 2026-03-10T12:37:22.064465+0000 mgr.a (mgr.14150) 202 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:25.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:25 vm05 bash[29416]: cluster 2026-03-10T12:37:24.064815+0000 mgr.a (mgr.14150) 203 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:25.424 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:25 vm05 bash[29416]: cluster 2026-03-10T12:37:24.064815+0000 mgr.a (mgr.14150) 203 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:25.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:25 vm01 bash[35029]: cluster 2026-03-10T12:37:24.064815+0000 mgr.a (mgr.14150) 203 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:25.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:25 vm01 bash[35029]: cluster 2026-03-10T12:37:24.064815+0000 mgr.a (mgr.14150) 203 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:27.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:27 vm01 bash[35029]: cluster 2026-03-10T12:37:26.065073+0000 mgr.a (mgr.14150) 204 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:27.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:27 vm01 bash[35029]: cluster 2026-03-10T12:37:26.065073+0000 mgr.a (mgr.14150) 204 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:27.593 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:27 vm05 bash[29416]: cluster 2026-03-10T12:37:26.065073+0000 mgr.a (mgr.14150) 204 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:27.593 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:27 vm05 bash[29416]: cluster 2026-03-10T12:37:26.065073+0000 mgr.a (mgr.14150) 204 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:28.505 INFO:teuthology.orchestra.run.vm01.stderr:+ grep '\-\-cpus=2' /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/unit.run 2026-03-10T12:37:28.506 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-a --pids-limit=0 --cpus=2 -e CONTAINER_IMAGE=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc -e NODE_NAME=vm01 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a:/var/run/ceph:z -v /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a:/var/log/ceph:z -v /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config:/etc/ceph/ceph.conf:z quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true '--default-log-stderr-prefix=debug ' --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 2026-03-10T12:37:28.506 INFO:teuthology.orchestra.run.vm01.stderr:+ grep '\-\-debug_ms 10' /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/unit.run 2026-03-10T12:37:28.507 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-a --pids-limit=0 --cpus=2 -e CONTAINER_IMAGE=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc -e NODE_NAME=vm01 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a:/var/run/ceph:z -v /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a:/var/log/ceph:z -v /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config:/etc/ceph/ceph.conf:z quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true '--default-log-stderr-prefix=debug ' --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 2026-03-10T12:37:28.507 INFO:teuthology.orchestra.run.vm01.stderr:+ grep testing_custom_containers /etc/cephadm_testing/testing.txt 2026-03-10T12:37:28.508 INFO:teuthology.orchestra.run.vm01.stdout:testing_custom_containers 2026-03-10T12:37:28.508 INFO:teuthology.orchestra.run.vm01.stderr:++ find /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a -maxdepth 1 -type d -name 'container.bar.*' 2026-03-10T12:37:28.510 INFO:teuthology.orchestra.run.vm01.stderr:+ dir=/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/container.bar.vm01 2026-03-10T12:37:28.510 INFO:teuthology.orchestra.run.vm01.stderr:+ test -n /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/container.bar.vm01 2026-03-10T12:37:28.510 INFO:teuthology.orchestra.run.vm01.stderr:+ grep ok /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/container.bar.vm01/data/primary.txt 2026-03-10T12:37:28.510 INFO:teuthology.orchestra.run.vm01.stdout:ok 2026-03-10T12:37:28.510 INFO:teuthology.orchestra.run.vm01.stderr:+ grep from=centos /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/container.bar.vm01/data/from.txt 2026-03-10T12:37:28.511 INFO:teuthology.orchestra.run.vm01.stdout:from=centos 2026-03-10T12:37:28.512 INFO:teuthology.orchestra.run.vm01.stderr:+ test -s /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/container.bar.vm01/data/presized.dat 2026-03-10T12:37:28.513 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-10T12:37:28.516 INFO:tasks.cephadm:Teardown begin 2026-03-10T12:37:28.517 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:37:28.568 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:37:28.583 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-10T12:37:28.583 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f864ef90-1c7c-11f1-946a-0b343fff599a -- ceph mgr module disable cephadm 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.504220+0000 mon.a (mon.0) 72 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.504220+0000 mon.a (mon.0) 72 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.552217+0000 mon.a (mon.0) 73 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.552217+0000 mon.a (mon.0) 73 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.579131+0000 mon.a (mon.0) 74 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.579131+0000 mon.a (mon.0) 74 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.591699+0000 mon.a (mon.0) 75 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.591699+0000 mon.a (mon.0) 75 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.598238+0000 mon.a (mon.0) 76 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.598238+0000 mon.a (mon.0) 76 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.602890+0000 mon.a (mon.0) 77 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:28 vm05 bash[29416]: audit 2026-03-10T12:37:27.602890+0000 mon.a (mon.0) 77 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.504220+0000 mon.a (mon.0) 72 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.504220+0000 mon.a (mon.0) 72 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.552217+0000 mon.a (mon.0) 73 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.552217+0000 mon.a (mon.0) 73 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.579131+0000 mon.a (mon.0) 74 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.579131+0000 mon.a (mon.0) 74 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.591699+0000 mon.a (mon.0) 75 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.591699+0000 mon.a (mon.0) 75 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.598238+0000 mon.a (mon.0) 76 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.598238+0000 mon.a (mon.0) 76 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.602890+0000 mon.a (mon.0) 77 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:28.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:28 vm01 bash[35029]: audit 2026-03-10T12:37:27.602890+0000 mon.a (mon.0) 77 : audit [INF] from='mgr.14150 192.168.123.101:0/2381363097' entity='mgr.a' 2026-03-10T12:37:29.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:29 vm05 bash[29416]: cluster 2026-03-10T12:37:28.065397+0000 mgr.a (mgr.14150) 205 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:29.924 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:29 vm05 bash[29416]: cluster 2026-03-10T12:37:28.065397+0000 mgr.a (mgr.14150) 205 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:29.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:29 vm01 bash[35029]: cluster 2026-03-10T12:37:28.065397+0000 mgr.a (mgr.14150) 205 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:29.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:29 vm01 bash[35029]: cluster 2026-03-10T12:37:28.065397+0000 mgr.a (mgr.14150) 205 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:32.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:31 vm05 bash[29416]: cluster 2026-03-10T12:37:30.065681+0000 mgr.a (mgr.14150) 206 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:32.174 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:31 vm05 bash[29416]: cluster 2026-03-10T12:37:30.065681+0000 mgr.a (mgr.14150) 206 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:32.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:31 vm01 bash[35029]: cluster 2026-03-10T12:37:30.065681+0000 mgr.a (mgr.14150) 206 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:32.198 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:31 vm01 bash[35029]: cluster 2026-03-10T12:37:30.065681+0000 mgr.a (mgr.14150) 206 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:33.242 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/mon.a/config 2026-03-10T12:37:33.915 INFO:teuthology.orchestra.run.vm01.stderr:2026-03-10T12:37:33.913+0000 7f390ac45640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-10T12:37:33.915 INFO:teuthology.orchestra.run.vm01.stderr:2026-03-10T12:37:33.913+0000 7f390ac45640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-10T12:37:33.915 INFO:teuthology.orchestra.run.vm01.stderr:2026-03-10T12:37:33.913+0000 7f390ac45640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-10T12:37:33.916 INFO:teuthology.orchestra.run.vm01.stderr:2026-03-10T12:37:33.913+0000 7f390ac45640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-10T12:37:33.916 INFO:teuthology.orchestra.run.vm01.stderr:2026-03-10T12:37:33.913+0000 7f390ac45640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-10T12:37:33.916 INFO:teuthology.orchestra.run.vm01.stderr:2026-03-10T12:37:33.913+0000 7f390ac45640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-10T12:37:33.916 INFO:teuthology.orchestra.run.vm01.stderr:2026-03-10T12:37:33.913+0000 7f390ac45640 -1 monclient: keyring not found 2026-03-10T12:37:33.916 INFO:teuthology.orchestra.run.vm01.stderr:[errno 21] error connecting to the cluster 2026-03-10T12:37:34.012 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:37:34.012 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-10T12:37:34.012 DEBUG:teuthology.orchestra.run.vm01:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T12:37:34.015 DEBUG:teuthology.orchestra.run.vm05:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T12:37:34.039 INFO:tasks.cephadm:Stopping all daemons... 2026-03-10T12:37:34.039 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-10T12:37:34.039 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a 2026-03-10T12:37:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:33 vm01 bash[35029]: cluster 2026-03-10T12:37:32.065947+0000 mgr.a (mgr.14150) 207 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:33 vm01 bash[35029]: cluster 2026-03-10T12:37:32.065947+0000 mgr.a (mgr.14150) 207 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:34 vm01 systemd[1]: Stopping Ceph mon.a for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:37:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:34 vm01 bash[35029]: debug 2026-03-10T12:37:34.137+0000 7f841b053640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 (PID: 1) UID: 0 2026-03-10T12:37:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 12:37:34 vm01 bash[35029]: debug 2026-03-10T12:37:34.137+0000 7f841b053640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-10T12:37:34.402 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.a.service' 2026-03-10T12:37:34.414 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T12:37:34.414 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-10T12:37:34.414 INFO:tasks.cephadm.mon.b:Stopping mon.b... 2026-03-10T12:37:34.414 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.b 2026-03-10T12:37:34.421 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:33 vm05 bash[29416]: cluster 2026-03-10T12:37:32.065947+0000 mgr.a (mgr.14150) 207 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:34.425 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:33 vm05 bash[29416]: cluster 2026-03-10T12:37:32.065947+0000 mgr.a (mgr.14150) 207 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:37:34.672 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:34 vm05 systemd[1]: Stopping Ceph mon.b for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:37:34.672 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:34 vm05 bash[29416]: debug 2026-03-10T12:37:34.483+0000 7f1e4aad6640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 (PID: 1) UID: 0 2026-03-10T12:37:34.672 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:34 vm05 bash[29416]: debug 2026-03-10T12:37:34.483+0000 7f1e4aad6640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-10T12:37:34.672 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 10 12:37:34 vm05 bash[30676]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-mon-b 2026-03-10T12:37:34.682 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mon.b.service' 2026-03-10T12:37:34.734 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T12:37:34.734 INFO:tasks.cephadm.mon.b:Stopped mon.b 2026-03-10T12:37:34.734 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-10T12:37:34.734 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.a 2026-03-10T12:37:34.933 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.a.service' 2026-03-10T12:37:34.945 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T12:37:34.945 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-10T12:37:34.945 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-10T12:37:34.945 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.b 2026-03-10T12:37:35.124 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@mgr.b.service' 2026-03-10T12:37:35.137 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T12:37:35.137 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-10T12:37:35.137 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-10T12:37:35.137 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@osd.0 2026-03-10T12:37:35.448 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:37:35 vm01 systemd[1]: Stopping Ceph osd.0 for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:37:35.448 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:37:35 vm01 bash[29119]: debug 2026-03-10T12:37:35.245+0000 7f495fb03640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-10T12:37:35.448 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:37:35 vm01 bash[29119]: debug 2026-03-10T12:37:35.245+0000 7f495fb03640 -1 osd.0 14 *** Got signal Terminated *** 2026-03-10T12:37:35.448 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:37:35 vm01 bash[29119]: debug 2026-03-10T12:37:35.245+0000 7f495fb03640 -1 osd.0 14 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T12:37:40.699 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:37:40 vm01 bash[36965]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-osd-0 2026-03-10T12:37:40.699 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 12:37:40 vm01 bash[37024]: Error response from daemon: No such container: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-osd-0 2026-03-10T12:37:41.020 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@osd.0.service' 2026-03-10T12:37:41.047 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T12:37:41.047 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-10T12:37:41.047 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-10T12:37:41.047 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@osd.1 2026-03-10T12:37:41.424 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:37:41 vm05 systemd[1]: Stopping Ceph osd.1 for f864ef90-1c7c-11f1-946a-0b343fff599a... 2026-03-10T12:37:41.424 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:37:41 vm05 bash[25830]: debug 2026-03-10T12:37:41.111+0000 7fdd21936640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-10T12:37:41.424 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:37:41 vm05 bash[25830]: debug 2026-03-10T12:37:41.111+0000 7fdd21936640 -1 osd.1 14 *** Got signal Terminated *** 2026-03-10T12:37:41.424 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:37:41 vm05 bash[25830]: debug 2026-03-10T12:37:41.111+0000 7fdd21936640 -1 osd.1 14 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T12:37:46.740 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:37:46 vm05 bash[31052]: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-osd-1 2026-03-10T12:37:46.740 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 10 12:37:46 vm05 bash[31164]: Error response from daemon: No such container: ceph-f864ef90-1c7c-11f1-946a-0b343fff599a-osd-1 2026-03-10T12:37:47.087 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-f864ef90-1c7c-11f1-946a-0b343fff599a@osd.1.service' 2026-03-10T12:37:47.119 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T12:37:47.119 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-10T12:37:47.119 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid f864ef90-1c7c-11f1-946a-0b343fff599a --force --keep-logs 2026-03-10T12:37:47.211 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:37:54.195 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid f864ef90-1c7c-11f1-946a-0b343fff599a --force --keep-logs 2026-03-10T12:37:54.287 INFO:teuthology.orchestra.run.vm05.stdout:Deleting cluster with fsid: f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:38:01.557 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:38:01.565 INFO:teuthology.orchestra.run.vm01.stderr:rm: cannot remove '/etc/ceph/ceph.client.admin.keyring': Is a directory 2026-03-10T12:38:01.565 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:38:01.566 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:38:01.574 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-10T12:38:01.574 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028/remote/vm01/crash 2026-03-10T12:38:01.574 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/crash -- . 2026-03-10T12:38:01.614 INFO:teuthology.orchestra.run.vm01.stderr:tar: /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/crash: Cannot open: No such file or directory 2026-03-10T12:38:01.614 INFO:teuthology.orchestra.run.vm01.stderr:tar: Error is not recoverable: exiting now 2026-03-10T12:38:01.615 DEBUG:teuthology.misc:Transferring archived files from vm05:/var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028/remote/vm05/crash 2026-03-10T12:38:01.615 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/crash -- . 2026-03-10T12:38:01.627 INFO:teuthology.orchestra.run.vm05.stderr:tar: /var/lib/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/crash: Cannot open: No such file or directory 2026-03-10T12:38:01.627 INFO:teuthology.orchestra.run.vm05.stderr:tar: Error is not recoverable: exiting now 2026-03-10T12:38:01.627 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-10T12:38:01.628 DEBUG:teuthology.orchestra.run.vm01:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-10T12:38:01.670 INFO:tasks.cephadm:Compressing logs... 2026-03-10T12:38:01.670 DEBUG:teuthology.orchestra.run.vm01:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T12:38:01.715 DEBUG:teuthology.orchestra.run.vm05:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T12:38:01.725 INFO:teuthology.orchestra.run.vm05.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T12:38:01.726 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T12:38:01.726 INFO:teuthology.orchestra.run.vm01.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T12:38:01.726 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mgr.a.log 2026-03-10T12:38:01.726 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.log 2026-03-10T12:38:01.727 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T12:38:01.727 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.log 2026-03-10T12:38:01.728 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mon.b.log 2026-03-10T12:38:01.728 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mgr.a.log: 87.7% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T12:38:01.728 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.log: 87.0% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.log.gz 2026-03-10T12:38:01.728 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mon.a.log 2026-03-10T12:38:01.729 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-osd.1.log 2026-03-10T12:38:01.729 INFO:teuthology.orchestra.run.vm05.stderr: 85.6% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T12:38:01.729 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mgr.b.log 2026-03-10T12:38:01.729 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.log: 86.9% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.log.gz 2026-03-10T12:38:01.729 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.audit.log 2026-03-10T12:38:01.729 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-osd.1.log: /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mon.b.log: gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.audit.log 2026-03-10T12:38:01.738 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mon.a.log: gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-volume.log 2026-03-10T12:38:01.739 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mgr.b.log: 90.9% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mgr.b.log.gz 2026-03-10T12:38:01.739 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.audit.log: 90.1% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.audit.log.gz 2026-03-10T12:38:01.739 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.cephadm.log 2026-03-10T12:38:01.739 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-volume.log 2026-03-10T12:38:01.740 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.audit.log: 90.5% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.audit.log.gz 2026-03-10T12:38:01.743 INFO:teuthology.orchestra.run.vm05.stderr: 94.1% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-osd.1.log.gz 2026-03-10T12:38:01.744 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.cephadm.log 2026-03-10T12:38:01.749 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-volume.log: /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.cephadm.log: 79.2% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.cephadm.log.gz 2026-03-10T12:38:01.750 INFO:teuthology.orchestra.run.vm05.stderr: 95.9% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-volume.log.gz 2026-03-10T12:38:01.750 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-volume.log: 96.0% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-volume.log.gz 2026-03-10T12:38:01.750 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-osd.0.log 2026-03-10T12:38:01.750 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.cephadm.log: 80.4% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph.cephadm.log.gz 2026-03-10T12:38:01.774 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-osd.0.log: 94.1% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-osd.0.log.gz 2026-03-10T12:38:01.783 INFO:teuthology.orchestra.run.vm05.stderr: 93.5% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mon.b.log.gz 2026-03-10T12:38:01.784 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T12:38:01.784 INFO:teuthology.orchestra.run.vm05.stderr:real 0m0.065s 2026-03-10T12:38:01.784 INFO:teuthology.orchestra.run.vm05.stderr:user 0m0.079s 2026-03-10T12:38:01.784 INFO:teuthology.orchestra.run.vm05.stderr:sys 0m0.011s 2026-03-10T12:38:01.788 INFO:teuthology.orchestra.run.vm01.stderr: 88.8% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mgr.a.log.gz 2026-03-10T12:38:01.847 INFO:teuthology.orchestra.run.vm01.stderr: 92.0% -- replaced with /var/log/ceph/f864ef90-1c7c-11f1-946a-0b343fff599a/ceph-mon.a.log.gz 2026-03-10T12:38:01.849 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T12:38:01.849 INFO:teuthology.orchestra.run.vm01.stderr:real 0m0.132s 2026-03-10T12:38:01.849 INFO:teuthology.orchestra.run.vm01.stderr:user 0m0.167s 2026-03-10T12:38:01.849 INFO:teuthology.orchestra.run.vm01.stderr:sys 0m0.027s 2026-03-10T12:38:01.849 INFO:tasks.cephadm:Archiving logs... 2026-03-10T12:38:01.849 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028/remote/vm01/log 2026-03-10T12:38:01.849 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T12:38:01.916 DEBUG:teuthology.misc:Transferring archived files from vm05:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028/remote/vm05/log 2026-03-10T12:38:01.916 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T12:38:01.930 INFO:tasks.cephadm:Removing cluster... 2026-03-10T12:38:01.930 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid f864ef90-1c7c-11f1-946a-0b343fff599a --force 2026-03-10T12:38:02.046 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:38:03.314 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid f864ef90-1c7c-11f1-946a-0b343fff599a --force 2026-03-10T12:38:03.409 INFO:teuthology.orchestra.run.vm05.stdout:Deleting cluster with fsid: f864ef90-1c7c-11f1-946a-0b343fff599a 2026-03-10T12:38:04.673 INFO:tasks.cephadm:Removing cephadm ... 2026-03-10T12:38:04.673 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T12:38:04.677 DEBUG:teuthology.orchestra.run.vm05:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T12:38:04.679 INFO:tasks.cephadm:Teardown complete 2026-03-10T12:38:04.683 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-10T12:38:04.689 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-10T12:38:04.689 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T12:38:04.719 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T12:38:04.739 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-10T12:38:04.739 DEBUG:teuthology.orchestra.run.vm01:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-10T12:38:04.746 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-10T12:38:04.746 DEBUG:teuthology.orchestra.run.vm05:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-10T12:38:04.809 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:04.814 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:05.064 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:05.065 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:05.066 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:05.067 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:05.338 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:05.339 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:38:05.340 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T12:38:05.340 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:05.341 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:05.342 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:38:05.343 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T12:38:05.343 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:05.361 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:05.363 INFO:teuthology.orchestra.run.vm05.stdout: ceph* 2026-03-10T12:38:05.365 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:05.366 INFO:teuthology.orchestra.run.vm01.stdout: ceph* 2026-03-10T12:38:05.565 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T12:38:05.565 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-10T12:38:05.581 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T12:38:05.581 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-10T12:38:05.618 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118605 files and directories currently installed.) 2026-03-10T12:38:05.620 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:05.629 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118605 files and directories currently installed.) 2026-03-10T12:38:05.631 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:06.871 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:06.911 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:06.961 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:06.998 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:07.085 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:07.085 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:07.231 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:07.232 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:07.330 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:07.331 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:38:07.332 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T12:38:07.332 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:07.354 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:07.356 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm* cephadm* 2026-03-10T12:38:07.453 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:07.453 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:38:07.454 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T12:38:07.454 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:07.476 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:07.478 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm* cephadm* 2026-03-10T12:38:07.558 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T12:38:07.558 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 1775 kB disk space will be freed. 2026-03-10T12:38:07.609 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118603 files and directories currently installed.) 2026-03-10T12:38:07.612 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:07.725 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T12:38:07.725 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 1775 kB disk space will be freed. 2026-03-10T12:38:07.820 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118603 files and directories currently installed.) 2026-03-10T12:38:07.821 INFO:teuthology.orchestra.run.vm01.stdout:Removing cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:07.823 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:07.872 INFO:teuthology.orchestra.run.vm01.stdout:Looking for files to backup/remove ... 2026-03-10T12:38:07.873 INFO:teuthology.orchestra.run.vm01.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-10T12:38:07.875 INFO:teuthology.orchestra.run.vm01.stdout:Removing user `cephadm' ... 2026-03-10T12:38:07.875 INFO:teuthology.orchestra.run.vm01.stdout:Warning: group `nogroup' has no more members. 2026-03-10T12:38:07.913 INFO:teuthology.orchestra.run.vm01.stdout:Done. 2026-03-10T12:38:07.917 INFO:teuthology.orchestra.run.vm05.stdout:Removing cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:07.990 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:08.000 INFO:teuthology.orchestra.run.vm05.stdout:Looking for files to backup/remove ... 2026-03-10T12:38:08.001 INFO:teuthology.orchestra.run.vm05.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-10T12:38:08.003 INFO:teuthology.orchestra.run.vm05.stdout:Removing user `cephadm' ... 2026-03-10T12:38:08.003 INFO:teuthology.orchestra.run.vm05.stdout:Warning: group `nogroup' has no more members. 2026-03-10T12:38:08.039 INFO:teuthology.orchestra.run.vm05.stdout:Done. 2026-03-10T12:38:08.083 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:08.330 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T12:38:08.332 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:08.441 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T12:38:08.445 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:09.643 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:09.679 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:09.683 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:09.723 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:09.919 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:09.919 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:09.985 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:09.985 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:10.099 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:10.099 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:38:10.099 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T12:38:10.099 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:10.114 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:10.115 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds* 2026-03-10T12:38:10.180 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:10.180 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T12:38:10.181 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T12:38:10.181 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:10.195 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:10.196 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds* 2026-03-10T12:38:10.309 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T12:38:10.309 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 7437 kB disk space will be freed. 2026-03-10T12:38:10.357 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T12:38:10.360 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:10.420 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T12:38:10.420 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 7437 kB disk space will be freed. 2026-03-10T12:38:10.457 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T12:38:10.459 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:10.831 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:10.928 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:11.021 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T12:38:11.023 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:11.137 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T12:38:11.139 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:12.708 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:12.746 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:12.818 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:12.857 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:13.042 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:13.042 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:13.051 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:13.052 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:13.168 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:13.169 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools python3-cheroot 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-psutil python3-pyinotify 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-threadpoolctl python3-waitress python3-webob python3-websocket 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev 2026-03-10T12:38:13.170 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:13.183 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:13.183 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-10T12:38:13.184 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-k8sevents* 2026-03-10T12:38:13.269 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools python3-cheroot 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-psutil python3-pyinotify 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-threadpoolctl python3-waitress python3-webob python3-websocket 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout: sg3-utils-udev 2026-03-10T12:38:13.270 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:13.284 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:13.284 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-10T12:38:13.286 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents* 2026-03-10T12:38:13.381 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 4 to remove and 10 not upgraded. 2026-03-10T12:38:13.381 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 165 MB disk space will be freed. 2026-03-10T12:38:13.423 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T12:38:13.426 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:13.447 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:13.501 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 4 to remove and 10 not upgraded. 2026-03-10T12:38:13.501 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 165 MB disk space will be freed. 2026-03-10T12:38:13.513 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:13.549 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T12:38:13.553 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:13.572 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:13.577 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:13.643 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:13.695 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:14.122 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T12:38:14.124 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:14.211 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T12:38:14.213 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:15.700 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:15.736 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:15.888 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:15.925 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:15.957 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:15.958 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:16.067 INFO:teuthology.orchestra.run.vm01.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:16.068 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:16.076 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:16.078 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-10T12:38:16.153 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:16.153 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:16.255 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T12:38:16.255 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 472 MB disk space will be freed. 2026-03-10T12:38:16.290 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T12:38:16.292 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:16.395 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:16.434 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:16.434 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:16.434 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:16.435 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:16.435 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:16.435 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:16.435 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:16.436 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:16.454 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:16.465 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-10T12:38:16.681 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T12:38:16.681 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 472 MB disk space will be freed. 2026-03-10T12:38:16.744 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T12:38:16.746 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:16.840 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:16.862 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:17.318 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:17.322 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:17.786 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:17.801 INFO:teuthology.orchestra.run.vm01.stdout:Removing radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:18.247 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:18.262 INFO:teuthology.orchestra.run.vm05.stdout:Removing radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:18.309 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:18.753 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:18.830 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:18.861 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:18.935 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:38:19.062 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117455 files and directories currently installed.) 2026-03-10T12:38:19.064 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:19.353 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:19.462 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:38:19.536 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117455 files and directories currently installed.) 2026-03-10T12:38:19.538 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:19.747 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:20.218 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:20.224 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:20.684 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:20.692 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:21.107 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:21.167 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:21.556 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:22.722 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:22.757 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:22.984 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:22.985 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:23.174 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:23.174 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:23.174 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:23.175 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:23.182 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:23.182 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse* 2026-03-10T12:38:23.356 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T12:38:23.356 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 3673 kB disk space will be freed. 2026-03-10T12:38:23.380 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:23.395 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117443 files and directories currently installed.) 2026-03-10T12:38:23.398 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:23.417 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:23.638 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:23.638 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:23.897 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:23.936 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:23.936 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:23.937 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:23.938 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:23.953 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:23.954 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse* 2026-03-10T12:38:24.101 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T12:38:24.104 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:24.142 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T12:38:24.142 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 3673 kB disk space will be freed. 2026-03-10T12:38:24.192 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117443 files and directories currently installed.) 2026-03-10T12:38:24.196 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:24.676 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:24.969 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T12:38:24.972 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:25.944 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:25.979 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:26.177 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:26.177 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:26.415 INFO:teuthology.orchestra.run.vm01.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-10T12:38:26.415 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:26.415 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:26.415 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:26.416 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:26.441 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:26.441 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:26.476 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:26.691 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:26.709 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:26.709 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:26.726 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:26.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:26.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:26.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:26.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:26.953 INFO:teuthology.orchestra.run.vm01.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:26.953 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:26.956 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:26.957 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:26.984 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:26.984 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:27.018 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:27.208 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:27.227 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:27.234 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:27.239 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:27.239 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:27.259 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:27.499 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:27.500 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:27.502 INFO:teuthology.orchestra.run.vm01.stdout:Package 'radosgw' is not installed, so not removed 2026-03-10T12:38:27.502 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:27.502 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:27.503 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:27.504 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:27.538 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:27.538 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:27.573 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:27.745 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:27.769 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:27.770 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:27.803 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:27.811 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:27.811 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:28.037 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:28.037 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:28.109 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:28.110 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:28.110 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:28.110 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet zip 2026-03-10T12:38:28.112 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:28.134 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:28.134 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout:Package 'radosgw' is not installed, so not removed 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T12:38:28.294 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T12:38:28.295 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T12:38:28.295 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T12:38:28.295 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:28.324 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:28.324 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:28.342 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 3 to remove and 10 not upgraded. 2026-03-10T12:38:28.342 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 2062 kB disk space will be freed. 2026-03-10T12:38:28.357 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:28.392 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T12:38:28.395 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:28.408 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:28.472 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:28.604 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:28.605 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-10T12:38:28.862 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:28.875 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:28.876 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-10T12:38:29.054 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 3 to remove and 10 not upgraded. 2026-03-10T12:38:29.054 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 2062 kB disk space will be freed. 2026-03-10T12:38:29.098 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T12:38:29.101 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:29.170 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:29.189 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:29.748 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:29.781 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:30.012 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:30.013 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:30.196 INFO:teuthology.orchestra.run.vm01.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-10T12:38:30.196 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:30.196 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:30.196 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:30.196 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet zip 2026-03-10T12:38:30.197 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:30.227 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:30.227 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:30.261 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:30.458 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:30.459 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:30.504 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:30.540 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:30.640 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:30.640 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:30.666 INFO:teuthology.orchestra.run.vm01.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-10T12:38:30.666 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:30.666 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:30.666 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:30.666 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet zip 2026-03-10T12:38:30.667 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:30.693 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:30.693 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:30.727 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:30.837 INFO:teuthology.orchestra.run.vm05.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-10T12:38:30.837 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:30.837 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:30.837 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:30.837 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-10T12:38:30.838 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:30.860 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:30.860 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:30.891 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:30.938 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:30.938 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:31.100 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:31.101 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:31.149 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:31.149 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:31.149 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:31.150 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:31.151 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:31.151 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet zip 2026-03-10T12:38:31.151 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:31.166 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:31.166 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd* 2026-03-10T12:38:31.284 INFO:teuthology.orchestra.run.vm05.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-10T12:38:31.284 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:31.284 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:31.284 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:31.284 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-10T12:38:31.285 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:31.307 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:31.307 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:31.339 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:31.348 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T12:38:31.348 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 1186 kB disk space will be freed. 2026-03-10T12:38:31.386 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117410 files and directories currently installed.) 2026-03-10T12:38:31.388 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:31.548 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:31.548 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:31.747 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:31.747 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:31.747 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:31.747 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-10T12:38:31.748 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:31.760 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:31.760 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd* 2026-03-10T12:38:31.930 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T12:38:31.930 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 1186 kB disk space will be freed. 2026-03-10T12:38:31.969 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117410 files and directories currently installed.) 2026-03-10T12:38:31.972 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:32.505 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:32.539 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:32.745 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:32.746 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:32.937 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:32.937 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:32.937 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:32.937 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet zip 2026-03-10T12:38:32.938 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:32.948 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:32.948 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-dev* libcephfs2* 2026-03-10T12:38:32.994 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:33.028 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:33.108 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T12:38:33.108 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 3202 kB disk space will be freed. 2026-03-10T12:38:33.143 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117402 files and directories currently installed.) 2026-03-10T12:38:33.145 INFO:teuthology.orchestra.run.vm01.stdout:Removing libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:33.155 INFO:teuthology.orchestra.run.vm01.stdout:Removing libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:33.178 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:38:33.216 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:33.216 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:33.422 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:33.422 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:33.422 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:33.422 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-10T12:38:33.423 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:33.439 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:33.440 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-dev* libcephfs2* 2026-03-10T12:38:33.617 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T12:38:33.618 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 3202 kB disk space will be freed. 2026-03-10T12:38:33.659 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117402 files and directories currently installed.) 2026-03-10T12:38:33.661 INFO:teuthology.orchestra.run.vm05.stdout:Removing libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:33.674 INFO:teuthology.orchestra.run.vm05.stdout:Removing libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:33.700 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:38:34.289 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:34.323 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:34.527 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:34.528 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:34.688 INFO:teuthology.orchestra.run.vm01.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-10T12:38:34.688 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:34.688 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:34.688 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:34.688 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet zip 2026-03-10T12:38:34.689 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:34.713 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:34.713 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:34.746 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:34.843 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:34.877 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:34.936 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:34.937 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:35.070 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:35.071 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:35.071 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:35.071 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:35.071 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:35.071 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T12:38:35.071 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T12:38:35.071 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:35.077 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:35.078 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:35.083 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:35.084 INFO:teuthology.orchestra.run.vm01.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-10T12:38:35.084 INFO:teuthology.orchestra.run.vm01.stdout: qemu-block-extra* rbd-fuse* 2026-03-10T12:38:35.254 INFO:teuthology.orchestra.run.vm05.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-10T12:38:35.254 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:35.254 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:35.254 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T12:38:35.254 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T12:38:35.254 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T12:38:35.254 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:35.254 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-10T12:38:35.255 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:35.259 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T12:38:35.259 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 51.6 MB disk space will be freed. 2026-03-10T12:38:35.274 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:35.274 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:35.297 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117387 files and directories currently installed.) 2026-03-10T12:38:35.299 INFO:teuthology.orchestra.run.vm01.stdout:Removing rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.306 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:35.308 INFO:teuthology.orchestra.run.vm01.stdout:Removing libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.319 INFO:teuthology.orchestra.run.vm01.stdout:Removing libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.327 INFO:teuthology.orchestra.run.vm01.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T12:38:35.493 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:35.494 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:35.635 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:35.635 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:35.635 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T12:38:35.635 INFO:teuthology.orchestra.run.vm05.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:38:35.635 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T12:38:35.636 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:35.649 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:35.650 INFO:teuthology.orchestra.run.vm05.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-10T12:38:35.650 INFO:teuthology.orchestra.run.vm05.stdout: qemu-block-extra* rbd-fuse* 2026-03-10T12:38:35.751 INFO:teuthology.orchestra.run.vm01.stdout:Removing librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.762 INFO:teuthology.orchestra.run.vm01.stdout:Removing librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.775 INFO:teuthology.orchestra.run.vm01.stdout:Removing librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.801 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:35.825 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T12:38:35.825 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 51.6 MB disk space will be freed. 2026-03-10T12:38:35.833 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:38:35.862 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117387 files and directories currently installed.) 2026-03-10T12:38:35.865 INFO:teuthology.orchestra.run.vm05.stdout:Removing rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.879 INFO:teuthology.orchestra.run.vm05.stdout:Removing libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.894 INFO:teuthology.orchestra.run.vm05.stdout:Removing libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:35.905 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T12:38:35.905 INFO:teuthology.orchestra.run.vm05.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T12:38:35.907 INFO:teuthology.orchestra.run.vm01.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T12:38:36.294 INFO:teuthology.orchestra.run.vm05.stdout:Removing librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:36.308 INFO:teuthology.orchestra.run.vm05.stdout:Removing librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:36.325 INFO:teuthology.orchestra.run.vm05.stdout:Removing librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:36.353 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:36.398 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:38:36.471 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T12:38:36.473 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T12:38:37.403 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:37.437 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:37.603 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:37.604 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:37.814 INFO:teuthology.orchestra.run.vm01.stdout:Package 'librbd1' is not installed, so not removed 2026-03-10T12:38:37.814 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:37.814 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:37.814 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T12:38:37.814 INFO:teuthology.orchestra.run.vm01.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:38:37.814 INFO:teuthology.orchestra.run.vm01.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T12:38:37.814 INFO:teuthology.orchestra.run.vm01.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T12:38:37.815 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:37.841 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:37.841 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:37.873 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:38.009 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:38.042 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:38.063 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:38.063 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:38.207 INFO:teuthology.orchestra.run.vm01.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-10T12:38:38.207 INFO:teuthology.orchestra.run.vm01.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:38.207 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:38.208 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:38.209 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T12:38:38.209 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T12:38:38.209 INFO:teuthology.orchestra.run.vm01.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:38.231 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:38.231 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:38.232 DEBUG:teuthology.orchestra.run.vm01:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-10T12:38:38.254 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:38.254 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:38.287 DEBUG:teuthology.orchestra.run.vm01:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-10T12:38:38.363 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:38.428 INFO:teuthology.orchestra.run.vm05.stdout:Package 'librbd1' is not installed, so not removed 2026-03-10T12:38:38.428 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:38.428 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:38.428 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:38.429 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T12:38:38.430 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T12:38:38.430 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:38.452 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:38.452 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:38.484 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:38.572 INFO:teuthology.orchestra.run.vm01.stdout:Building dependency tree... 2026-03-10T12:38:38.573 INFO:teuthology.orchestra.run.vm01.stdout:Reading state information... 2026-03-10T12:38:38.703 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:38.704 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:38.790 INFO:teuthology.orchestra.run.vm01.stdout:The following packages will be REMOVED: 2026-03-10T12:38:38.790 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:38.790 INFO:teuthology.orchestra.run.vm01.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T12:38:38.790 INFO:teuthology.orchestra.run.vm01.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:38:38.790 INFO:teuthology.orchestra.run.vm01.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T12:38:38.790 INFO:teuthology.orchestra.run.vm01.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T12:38:38.791 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T12:38:38.870 INFO:teuthology.orchestra.run.vm05.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-10T12:38:38.870 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T12:38:38.870 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:38.870 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T12:38:38.870 INFO:teuthology.orchestra.run.vm05.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:38:38.870 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T12:38:38.870 INFO:teuthology.orchestra.run.vm05.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T12:38:38.871 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T12:38:38.895 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T12:38:38.895 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:38.896 DEBUG:teuthology.orchestra.run.vm05:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-10T12:38:38.950 DEBUG:teuthology.orchestra.run.vm05:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-10T12:38:38.975 INFO:teuthology.orchestra.run.vm01.stdout:0 upgraded, 0 newly installed, 87 to remove and 10 not upgraded. 2026-03-10T12:38:38.975 INFO:teuthology.orchestra.run.vm01.stdout:After this operation, 107 MB disk space will be freed. 2026-03-10T12:38:39.007 INFO:teuthology.orchestra.run.vm01.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T12:38:39.009 INFO:teuthology.orchestra.run.vm01.stdout:Removing ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:39.025 INFO:teuthology.orchestra.run.vm01.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-10T12:38:39.027 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:39.036 INFO:teuthology.orchestra.run.vm01.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-10T12:38:39.046 INFO:teuthology.orchestra.run.vm01.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T12:38:39.056 INFO:teuthology.orchestra.run.vm01.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T12:38:39.066 INFO:teuthology.orchestra.run.vm01.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T12:38:39.075 INFO:teuthology.orchestra.run.vm01.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:38:39.085 INFO:teuthology.orchestra.run.vm01.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:38:39.094 INFO:teuthology.orchestra.run.vm01.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:38:39.113 INFO:teuthology.orchestra.run.vm01.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T12:38:39.123 INFO:teuthology.orchestra.run.vm01.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T12:38:39.132 INFO:teuthology.orchestra.run.vm01.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T12:38:39.143 INFO:teuthology.orchestra.run.vm01.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T12:38:39.155 INFO:teuthology.orchestra.run.vm01.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T12:38:39.166 INFO:teuthology.orchestra.run.vm01.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T12:38:39.177 INFO:teuthology.orchestra.run.vm01.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-10T12:38:39.189 INFO:teuthology.orchestra.run.vm01.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T12:38:39.200 INFO:teuthology.orchestra.run.vm01.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T12:38:39.213 INFO:teuthology.orchestra.run.vm01.stdout:Removing luarocks (3.8.0+dfsg1-1) ... 2026-03-10T12:38:39.225 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-10T12:38:39.225 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-10T12:38:39.238 INFO:teuthology.orchestra.run.vm01.stdout:Removing liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T12:38:39.250 INFO:teuthology.orchestra.run.vm01.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-10T12:38:39.260 INFO:teuthology.orchestra.run.vm01.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T12:38:39.271 INFO:teuthology.orchestra.run.vm01.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T12:38:39.282 INFO:teuthology.orchestra.run.vm01.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T12:38:39.293 INFO:teuthology.orchestra.run.vm01.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-10T12:38:39.304 INFO:teuthology.orchestra.run.vm01.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T12:38:39.317 INFO:teuthology.orchestra.run.vm01.stdout:Removing libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T12:38:39.328 INFO:teuthology.orchestra.run.vm01.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-10T12:38:39.335 INFO:teuthology.orchestra.run.vm01.stdout:update-initramfs: deferring update (trigger activated) 2026-03-10T12:38:39.345 INFO:teuthology.orchestra.run.vm01.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-10T12:38:39.364 INFO:teuthology.orchestra.run.vm01.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-10T12:38:39.371 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-10T12:38:39.371 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T12:38:39.371 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T12:38:39.371 INFO:teuthology.orchestra.run.vm05.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T12:38:39.371 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T12:38:39.371 INFO:teuthology.orchestra.run.vm05.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T12:38:39.372 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T12:38:39.375 INFO:teuthology.orchestra.run.vm01.stdout:Removing lua-any (27ubuntu1) ... 2026-03-10T12:38:39.385 INFO:teuthology.orchestra.run.vm01.stdout:Removing lua-sec:amd64 (1.0.2-1) ... 2026-03-10T12:38:39.395 INFO:teuthology.orchestra.run.vm01.stdout:Removing lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T12:38:39.408 INFO:teuthology.orchestra.run.vm01.stdout:Removing lua5.1 (5.1.5-8.1build4) ... 2026-03-10T12:38:39.425 INFO:teuthology.orchestra.run.vm01.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T12:38:39.544 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 87 to remove and 10 not upgraded. 2026-03-10T12:38:39.544 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 107 MB disk space will be freed. 2026-03-10T12:38:39.577 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T12:38:39.578 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:39.593 INFO:teuthology.orchestra.run.vm05.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-10T12:38:39.604 INFO:teuthology.orchestra.run.vm05.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-10T12:38:39.614 INFO:teuthology.orchestra.run.vm05.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T12:38:39.626 INFO:teuthology.orchestra.run.vm05.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T12:38:39.638 INFO:teuthology.orchestra.run.vm05.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T12:38:39.650 INFO:teuthology.orchestra.run.vm05.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:38:39.663 INFO:teuthology.orchestra.run.vm05.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:38:39.675 INFO:teuthology.orchestra.run.vm05.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T12:38:39.694 INFO:teuthology.orchestra.run.vm05.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T12:38:39.704 INFO:teuthology.orchestra.run.vm05.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T12:38:39.715 INFO:teuthology.orchestra.run.vm05.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T12:38:39.725 INFO:teuthology.orchestra.run.vm05.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T12:38:39.736 INFO:teuthology.orchestra.run.vm05.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T12:38:39.747 INFO:teuthology.orchestra.run.vm05.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T12:38:39.758 INFO:teuthology.orchestra.run.vm05.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-10T12:38:39.768 INFO:teuthology.orchestra.run.vm05.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T12:38:39.778 INFO:teuthology.orchestra.run.vm05.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T12:38:39.789 INFO:teuthology.orchestra.run.vm05.stdout:Removing luarocks (3.8.0+dfsg1-1) ... 2026-03-10T12:38:39.814 INFO:teuthology.orchestra.run.vm05.stdout:Removing liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T12:38:39.826 INFO:teuthology.orchestra.run.vm05.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-10T12:38:39.828 INFO:teuthology.orchestra.run.vm01.stdout:Removing pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T12:38:39.836 INFO:teuthology.orchestra.run.vm05.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T12:38:39.846 INFO:teuthology.orchestra.run.vm05.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T12:38:39.857 INFO:teuthology.orchestra.run.vm05.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T12:38:39.857 INFO:teuthology.orchestra.run.vm01.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T12:38:39.868 INFO:teuthology.orchestra.run.vm05.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-10T12:38:39.880 INFO:teuthology.orchestra.run.vm05.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T12:38:39.885 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T12:38:39.899 INFO:teuthology.orchestra.run.vm05.stdout:Removing libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T12:38:39.910 INFO:teuthology.orchestra.run.vm05.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-10T12:38:39.927 INFO:teuthology.orchestra.run.vm05.stdout:update-initramfs: deferring update (trigger activated) 2026-03-10T12:38:39.967 INFO:teuthology.orchestra.run.vm05.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-10T12:38:39.974 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-webtest (2.0.35-1) ... 2026-03-10T12:38:39.985 INFO:teuthology.orchestra.run.vm05.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-10T12:38:39.996 INFO:teuthology.orchestra.run.vm05.stdout:Removing lua-any (27ubuntu1) ... 2026-03-10T12:38:40.008 INFO:teuthology.orchestra.run.vm05.stdout:Removing lua-sec:amd64 (1.0.2-1) ... 2026-03-10T12:38:40.019 INFO:teuthology.orchestra.run.vm05.stdout:Removing lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T12:38:40.023 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-pastescript (2.0.2-4) ... 2026-03-10T12:38:40.033 INFO:teuthology.orchestra.run.vm05.stdout:Removing lua5.1 (5.1.5-8.1build4) ... 2026-03-10T12:38:40.051 INFO:teuthology.orchestra.run.vm05.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T12:38:40.074 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-pastedeploy (2.1.1-1) ... 2026-03-10T12:38:40.118 INFO:teuthology.orchestra.run.vm01.stdout:Removing python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T12:38:40.128 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T12:38:40.303 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T12:38:40.673 INFO:teuthology.orchestra.run.vm05.stdout:Removing pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T12:38:40.717 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-10T12:38:40.724 INFO:teuthology.orchestra.run.vm05.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T12:38:40.879 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T12:38:40.898 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-10T12:38:40.999 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:41.021 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-webtest (2.0.35-1) ... 2026-03-10T12:38:41.061 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:41.081 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-pastescript (2.0.2-4) ... 2026-03-10T12:38:41.112 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-10T12:38:41.136 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-pastedeploy (2.1.1-1) ... 2026-03-10T12:38:41.171 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T12:38:41.184 INFO:teuthology.orchestra.run.vm05.stdout:Removing python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T12:38:41.198 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T12:38:41.221 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-10T12:38:41.257 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T12:38:41.267 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-10T12:38:41.314 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-10T12:38:41.570 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-10T12:38:41.594 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-10T12:38:41.616 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-10T12:38:41.643 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-10T12:38:41.663 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-10T12:38:41.691 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:41.713 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T12:38:41.741 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T12:38:41.790 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-10T12:38:41.840 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T12:38:41.848 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T12:38:41.897 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-10T12:38:41.898 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-logutils (0.3.3-8) ... 2026-03-10T12:38:41.943 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-10T12:38:41.948 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T12:38:41.991 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-10T12:38:41.999 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-10T12:38:42.038 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-10T12:38:42.049 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T12:38:42.084 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-10T12:38:42.106 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-10T12:38:42.130 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-10T12:38:42.153 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-10T12:38:42.176 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T12:38:42.203 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-pyinotify (0.9.6-1.3) ... 2026-03-10T12:38:42.251 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T12:38:42.299 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T12:38:42.302 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-10T12:38:42.351 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T12:38:42.358 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-logutils (0.3.3-8) ... 2026-03-10T12:38:42.401 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-rsa (4.8-1) ... 2026-03-10T12:38:42.407 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T12:38:42.455 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-simplegeneric (0.8.1-3) ... 2026-03-10T12:38:42.460 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-10T12:38:42.500 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-10T12:38:42.507 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T12:38:42.554 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-singledispatch (3.4.0.3-3) ... 2026-03-10T12:38:42.563 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-10T12:38:42.600 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T12:38:42.610 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-10T12:38:42.626 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T12:38:42.660 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-pyinotify (0.9.6-1.3) ... 2026-03-10T12:38:42.673 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-10T12:38:42.707 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T12:38:42.718 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T12:38:42.758 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-10T12:38:42.766 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T12:38:42.807 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T12:38:42.812 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T12:38:42.857 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-rsa (4.8-1) ... 2026-03-10T12:38:42.861 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-10T12:38:42.906 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-simplegeneric (0.8.1-3) ... 2026-03-10T12:38:42.908 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T12:38:42.952 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-10T12:38:42.958 INFO:teuthology.orchestra.run.vm01.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-10T12:38:43.005 INFO:teuthology.orchestra.run.vm01.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-10T12:38:43.005 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-singledispatch (3.4.0.3-3) ... 2026-03-10T12:38:43.027 INFO:teuthology.orchestra.run.vm01.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T12:38:43.056 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T12:38:43.083 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T12:38:43.129 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-10T12:38:43.175 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T12:38:43.222 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T12:38:43.268 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T12:38:43.319 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-10T12:38:43.369 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T12:38:43.403 INFO:teuthology.orchestra.run.vm01.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-10T12:38:43.415 INFO:teuthology.orchestra.run.vm01.stdout:Removing unzip (6.0-26ubuntu3.2) ... 2026-03-10T12:38:43.421 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-10T12:38:43.433 INFO:teuthology.orchestra.run.vm01.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-10T12:38:43.450 INFO:teuthology.orchestra.run.vm01.stdout:Removing zip (3.0-12build2) ... 2026-03-10T12:38:43.469 INFO:teuthology.orchestra.run.vm05.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-10T12:38:43.476 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:38:43.486 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:43.491 INFO:teuthology.orchestra.run.vm05.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T12:38:43.530 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T12:38:43.537 INFO:teuthology.orchestra.run.vm01.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-10T12:38:43.556 INFO:teuthology.orchestra.run.vm01.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-1092-kvm 2026-03-10T12:38:43.904 INFO:teuthology.orchestra.run.vm05.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-10T12:38:43.917 INFO:teuthology.orchestra.run.vm05.stdout:Removing unzip (6.0-26ubuntu3.2) ... 2026-03-10T12:38:43.938 INFO:teuthology.orchestra.run.vm05.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-10T12:38:43.955 INFO:teuthology.orchestra.run.vm05.stdout:Removing zip (3.0-12build2) ... 2026-03-10T12:38:43.980 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T12:38:43.990 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T12:38:44.032 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T12:38:44.040 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-10T12:38:44.058 INFO:teuthology.orchestra.run.vm05.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-1092-kvm 2026-03-10T12:38:45.042 INFO:teuthology.orchestra.run.vm01.stdout:W: mkconf: MD subsystem is not loaded, thus I cannot scan for arrays. 2026-03-10T12:38:45.044 INFO:teuthology.orchestra.run.vm01.stdout:W: mdadm: failed to auto-generate temporary mdadm.conf file. 2026-03-10T12:38:45.595 INFO:teuthology.orchestra.run.vm05.stdout:W: mkconf: MD subsystem is not loaded, thus I cannot scan for arrays. 2026-03-10T12:38:45.597 INFO:teuthology.orchestra.run.vm05.stdout:W: mdadm: failed to auto-generate temporary mdadm.conf file. 2026-03-10T12:38:47.174 INFO:teuthology.orchestra.run.vm01.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:47.177 DEBUG:teuthology.parallel:result is None 2026-03-10T12:38:47.950 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T12:38:47.953 DEBUG:teuthology.parallel:result is None 2026-03-10T12:38:47.953 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm01.local 2026-03-10T12:38:47.953 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm05.local 2026-03-10T12:38:47.953 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-10T12:38:47.953 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-10T12:38:47.960 DEBUG:teuthology.orchestra.run.vm01:> sudo apt-get update 2026-03-10T12:38:48.002 DEBUG:teuthology.orchestra.run.vm05:> sudo apt-get update 2026-03-10T12:38:48.145 INFO:teuthology.orchestra.run.vm01.stdout:Hit:1 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T12:38:48.148 INFO:teuthology.orchestra.run.vm01.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T12:38:48.159 INFO:teuthology.orchestra.run.vm01.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T12:38:48.185 INFO:teuthology.orchestra.run.vm05.stdout:Hit:1 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T12:38:48.192 INFO:teuthology.orchestra.run.vm05.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T12:38:48.200 INFO:teuthology.orchestra.run.vm05.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T12:38:48.208 INFO:teuthology.orchestra.run.vm05.stdout:Hit:4 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T12:38:48.253 INFO:teuthology.orchestra.run.vm01.stdout:Hit:4 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T12:38:49.163 INFO:teuthology.orchestra.run.vm01.stdout:Reading package lists... 2026-03-10T12:38:49.179 DEBUG:teuthology.parallel:result is None 2026-03-10T12:38:49.189 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-10T12:38:49.200 DEBUG:teuthology.parallel:result is None 2026-03-10T12:38:49.200 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-10T12:38:49.202 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-10T12:38:49.203 DEBUG:teuthology.orchestra.run.vm01:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T12:38:49.204 DEBUG:teuthology.orchestra.run.vm05:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:============================================================================== 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:#ec2-18-192-244- 216.239.35.8 2 u 68 64 377 23.542 -0.727 1.825 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:+mail.klausen.dk 193.79.237.14 2 u 62 64 177 23.610 +0.311 0.743 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:+sv1.ggsrv.de 192.53.103.103 2 u 56 64 377 24.959 +0.675 0.592 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:-ntp1.noris.net 182.107.54.146 2 u 54 64 377 23.590 +0.541 0.396 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:*ntp1.aew1.soe.a .GPS. 1 u 54 64 377 25.298 +0.536 0.245 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:#mail.morbitzer. 205.46.178.169 2 u 56 64 377 28.288 -2.691 0.797 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:-time.cloudflare 10.71.2.234 3 u 58 64 377 20.419 +1.647 0.386 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:-mail.gunnarhofm 192.53.103.103 2 u 52 64 377 25.037 +0.480 0.304 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:+obelix.hetzner. 77.90.0.148 3 u 55 64 377 25.034 +0.109 0.243 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:#82.165.178.31 82.64.45.50 2 u 54 64 377 27.205 -0.284 0.281 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:-time.cloudflare 10.17.8.4 3 u 51 64 377 20.420 +1.423 0.419 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:#v22025082392863 129.69.253.1 2 u 52 64 377 28.670 -2.509 0.373 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:#185.125.190.58 145.238.80.80 2 u 60 64 377 35.115 -1.133 1.522 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:#nox.spnr.de 205.46.178.169 2 u 52 64 377 38.548 -5.937 0.572 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:+static.46.170.2 188.40.142.18 3 u 54 64 377 25.010 +0.224 0.212 2026-03-10T12:38:50.104 INFO:teuthology.orchestra.run.vm05.stdout:-ntp2.wup-de.hos 237.17.204.95 2 u 57 64 377 33.672 +1.958 0.453 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:============================================================================== 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:*ntp1.aew1.soe.a .GPS. 1 u 58 64 377 25.283 +1.316 0.832 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:-obelix.hetzner. 77.90.0.148 3 u 56 64 377 25.044 +0.650 0.570 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:+sv1.ggsrv.de 192.53.103.103 2 u 63 64 377 24.925 +1.255 0.797 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:#ec2-18-192-244- 216.239.35.8 2 u - 64 377 23.800 +0.291 1.190 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:-mail.gunnarhofm 192.53.103.103 2 u 52 64 377 25.000 +0.815 0.616 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:#mail.morbitzer. 205.46.178.169 2 u 57 64 377 28.270 -2.050 0.969 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:-ntp1.noris.net 182.107.54.146 2 u 57 64 377 23.535 -0.120 1.281 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:+mail.klausen.dk 193.79.237.14 2 u 63 64 177 23.562 +0.962 0.583 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:#v22025082392863 129.69.253.1 2 u 54 64 377 28.216 -2.016 1.031 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:#ntp2.wup-de.hos 237.17.204.95 2 u 57 64 377 33.715 +2.447 0.646 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:-time.cloudflare 10.216.8.4 3 u 53 64 377 20.407 +1.023 0.884 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:+static.46.170.2 188.40.142.18 3 u 61 64 377 25.011 +0.953 0.606 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:-ntp5.kernfusion 237.17.204.95 2 u 57 64 377 28.960 +1.195 0.651 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:#185.125.190.58 145.238.80.80 2 u - 64 377 36.607 +0.046 0.929 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:-time.cloudflare 10.71.2.234 3 u 50 64 377 20.380 +2.418 0.740 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:#82.165.178.31 82.64.45.50 2 u 56 64 377 27.069 +0.529 0.657 2026-03-10T12:38:50.215 INFO:teuthology.orchestra.run.vm01.stdout:#alphyn.canonica 132.163.96.1 2 u 5 64 377 97.792 +0.435 1.780 2026-03-10T12:38:50.216 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-10T12:38:50.218 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-10T12:38:50.218 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-10T12:38:50.220 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-10T12:38:50.223 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-10T12:38:50.225 INFO:teuthology.task.internal:Duration was 674.977363 seconds 2026-03-10T12:38:50.225 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-10T12:38:50.227 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-10T12:38:50.227 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T12:38:50.228 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T12:38:50.250 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-10T12:38:50.251 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm01.local 2026-03-10T12:38:50.251 DEBUG:teuthology.orchestra.run.vm01:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T12:38:50.299 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm05.local 2026-03-10T12:38:50.299 DEBUG:teuthology.orchestra.run.vm05:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T12:38:50.309 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-10T12:38:50.309 DEBUG:teuthology.orchestra.run.vm01:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T12:38:50.342 DEBUG:teuthology.orchestra.run.vm05:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T12:38:50.394 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-10T12:38:50.394 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T12:38:50.426 DEBUG:teuthology.orchestra.run.vm05:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T12:38:50.432 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T12:38:50.432 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T12:38:50.432 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T12:38:50.432 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T12:38:50.432 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T12:38:50.441 INFO:teuthology.orchestra.run.vm01.stderr: 88.5% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T12:38:50.446 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T12:38:50.446 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T12:38:50.446 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T12:38:50.446 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T12:38:50.447 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T12:38:50.455 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 88.6% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T12:38:50.456 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-10T12:38:50.458 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-10T12:38:50.458 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T12:38:50.492 DEBUG:teuthology.orchestra.run.vm05:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T12:38:50.506 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-10T12:38:50.510 DEBUG:teuthology.orchestra.run.vm01:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T12:38:50.534 DEBUG:teuthology.orchestra.run.vm05:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T12:38:50.540 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = core 2026-03-10T12:38:50.557 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = core 2026-03-10T12:38:50.565 DEBUG:teuthology.orchestra.run.vm01:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T12:38:50.592 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:38:50.592 DEBUG:teuthology.orchestra.run.vm05:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T12:38:50.609 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:38:50.609 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-10T12:38:50.611 INFO:teuthology.task.internal:Transferring archived files... 2026-03-10T12:38:50.612 DEBUG:teuthology.misc:Transferring archived files from vm01:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028/remote/vm01 2026-03-10T12:38:50.612 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T12:38:50.642 DEBUG:teuthology.misc:Transferring archived files from vm05:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1028/remote/vm05 2026-03-10T12:38:50.642 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T12:38:50.658 INFO:teuthology.task.internal:Removing archive directory... 2026-03-10T12:38:50.658 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T12:38:50.686 DEBUG:teuthology.orchestra.run.vm05:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T12:38:50.706 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-10T12:38:50.709 INFO:teuthology.task.internal:Not uploading archives. 2026-03-10T12:38:50.709 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-10T12:38:50.712 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-10T12:38:50.712 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T12:38:50.730 DEBUG:teuthology.orchestra.run.vm05:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T12:38:50.732 INFO:teuthology.orchestra.run.vm01.stdout: 258075 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 10 12:38 /home/ubuntu/cephtest 2026-03-10T12:38:50.750 INFO:teuthology.orchestra.run.vm05.stdout: 258207 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 10 12:38 /home/ubuntu/cephtest 2026-03-10T12:38:50.750 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-10T12:38:50.758 INFO:teuthology.run:Summary data: description: orch/cephadm/workunits/{0-distro/ubuntu_22.04 agent/on mon_election/connectivity task/test_extra_daemon_features} duration: 674.97736287117 flavor: default owner: kyr success: true 2026-03-10T12:38:50.758 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T12:38:50.778 INFO:teuthology.run:pass