2026-03-08T23:30:31.263 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-08T23:30:31.268 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-08T23:30:31.291 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300 branch: squid description: orch:cephadm/workunits/{0-distro/ubuntu_22.04 agent/off mon_election/classic task/test_extra_daemon_features} email: null first_in_suite: false flavor: default job_id: '300' ktype: distro last_in_suite: false machine_type: vps name: kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps no_nested_subset: false os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 1 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: false mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath workunit: branch: tt-squid sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 seed: 8017 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch:cephadm suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 targets: vm02.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCE0id1IqtNu1ZnhaiWZ+gDOgPYF3qpB5XM5/fkhmJTplKDnPI5lRTaIlifc/Nsa2pie8tzC8mcgQ/ICECmpcQY= vm05.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGqo3JYENhEfRmfvvgFkMBMt0O1npOP2q9wEP7Wq3qonWuFOcQugtwIT6a90p+tE087nv2QiynwDPN8dVr5eZRw= tasks: - install: null - cephadm: null - exec: all-hosts: - mkdir /etc/cephadm_testing - cephadm.apply: specs: - extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon - custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest - custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z - cephadm.wait_for_service: service: mon - cephadm.wait_for_service: service: container.foo - cephadm.wait_for_service: service: container.bar - exec: host.a: - 'set -ex FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) sleep 60 # check extra container and entrypoint args written to mon unit run file grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run # check that custom container properly wrote content to file. # This requires the custom config, extra container args, and # entrypoint args to all be working in order for this to have # been written. The container entrypoint was set up with custom_configs, # the content and where to write to with the entrypoint args, and the mounting # of the /etc/cephadm_testing dir with extra container args grep "testing_custom_containers" /etc/cephadm_testing/testing.txt # Verify that container bar''s init containers and primary container # ran successfully dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name ''container.bar.*'') test -n "$dir" grep ok ${dir}/data/primary.txt grep from=centos ${dir}/data/from.txt test -s ${dir}/data/presized.dat ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-08_22:22:45 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-08T23:30:31.291 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa; will attempt to use it 2026-03-08T23:30:31.291 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa/tasks 2026-03-08T23:30:31.291 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-08T23:30:31.292 INFO:teuthology.task.internal:Checking packages... 2026-03-08T23:30:31.292 INFO:teuthology.task.internal:Checking packages for os_type 'ubuntu', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-08T23:30:31.292 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-08T23:30:31.292 INFO:teuthology.packaging:ref: None 2026-03-08T23:30:31.292 INFO:teuthology.packaging:tag: None 2026-03-08T23:30:31.292 INFO:teuthology.packaging:branch: squid 2026-03-08T23:30:31.292 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:30:31.292 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&ref=squid 2026-03-08T23:30:31.947 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:30:31.948 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-08T23:30:31.948 INFO:teuthology.task.internal:no buildpackages task found 2026-03-08T23:30:31.948 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-08T23:30:31.949 INFO:teuthology.task.internal:Saving configuration 2026-03-08T23:30:31.954 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-08T23:30:31.955 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-08T23:30:31.962 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm02.local', 'description': '/archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-08 23:29:22.476989', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:02', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCE0id1IqtNu1ZnhaiWZ+gDOgPYF3qpB5XM5/fkhmJTplKDnPI5lRTaIlifc/Nsa2pie8tzC8mcgQ/ICECmpcQY='} 2026-03-08T23:30:31.967 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm05.local', 'description': '/archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-08 23:29:22.476390', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:05', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGqo3JYENhEfRmfvvgFkMBMt0O1npOP2q9wEP7Wq3qonWuFOcQugtwIT6a90p+tE087nv2QiynwDPN8dVr5eZRw='} 2026-03-08T23:30:31.967 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-08T23:30:31.968 INFO:teuthology.task.internal:roles: ubuntu@vm02.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0'] 2026-03-08T23:30:31.968 INFO:teuthology.task.internal:roles: ubuntu@vm05.local - ['host.b', 'mon.b', 'mgr.b', 'osd.1'] 2026-03-08T23:30:31.968 INFO:teuthology.run_tasks:Running task console_log... 2026-03-08T23:30:31.974 DEBUG:teuthology.task.console_log:vm02 does not support IPMI; excluding 2026-03-08T23:30:31.980 DEBUG:teuthology.task.console_log:vm05 does not support IPMI; excluding 2026-03-08T23:30:31.980 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f39aac8caf0>, signals=[15]) 2026-03-08T23:30:31.980 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-08T23:30:31.980 INFO:teuthology.task.internal:Opening connections... 2026-03-08T23:30:31.980 DEBUG:teuthology.task.internal:connecting to ubuntu@vm02.local 2026-03-08T23:30:31.981 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-08T23:30:32.039 DEBUG:teuthology.task.internal:connecting to ubuntu@vm05.local 2026-03-08T23:30:32.040 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-08T23:30:32.101 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-08T23:30:32.102 DEBUG:teuthology.orchestra.run.vm02:> uname -m 2026-03-08T23:30:32.108 INFO:teuthology.orchestra.run.vm02.stdout:x86_64 2026-03-08T23:30:32.108 DEBUG:teuthology.orchestra.run.vm02:> cat /etc/os-release 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:NAME="Ubuntu" 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_ID="22.04" 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_CODENAME=jammy 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:ID=ubuntu 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:ID_LIKE=debian 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-08T23:30:32.155 INFO:teuthology.orchestra.run.vm02.stdout:UBUNTU_CODENAME=jammy 2026-03-08T23:30:32.155 INFO:teuthology.lock.ops:Updating vm02.local on lock server 2026-03-08T23:30:32.160 DEBUG:teuthology.orchestra.run.vm05:> uname -m 2026-03-08T23:30:32.181 INFO:teuthology.orchestra.run.vm05.stdout:x86_64 2026-03-08T23:30:32.181 DEBUG:teuthology.orchestra.run.vm05:> cat /etc/os-release 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:NAME="Ubuntu" 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:VERSION_ID="22.04" 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:VERSION_CODENAME=jammy 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:ID=ubuntu 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:ID_LIKE=debian 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-08T23:30:32.226 INFO:teuthology.orchestra.run.vm05.stdout:UBUNTU_CODENAME=jammy 2026-03-08T23:30:32.226 INFO:teuthology.lock.ops:Updating vm05.local on lock server 2026-03-08T23:30:32.230 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-08T23:30:32.232 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-08T23:30:32.233 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-08T23:30:32.233 DEBUG:teuthology.orchestra.run.vm02:> test '!' -e /home/ubuntu/cephtest 2026-03-08T23:30:32.234 DEBUG:teuthology.orchestra.run.vm05:> test '!' -e /home/ubuntu/cephtest 2026-03-08T23:30:32.270 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-08T23:30:32.271 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-08T23:30:32.271 DEBUG:teuthology.orchestra.run.vm02:> test -z $(ls -A /var/lib/ceph) 2026-03-08T23:30:32.277 DEBUG:teuthology.orchestra.run.vm05:> test -z $(ls -A /var/lib/ceph) 2026-03-08T23:30:32.279 INFO:teuthology.orchestra.run.vm02.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-08T23:30:32.314 INFO:teuthology.orchestra.run.vm05.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-08T23:30:32.315 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-08T23:30:32.322 DEBUG:teuthology.orchestra.run.vm02:> test -e /ceph-qa-ready 2026-03-08T23:30:32.325 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:30:32.653 DEBUG:teuthology.orchestra.run.vm05:> test -e /ceph-qa-ready 2026-03-08T23:30:32.655 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:30:32.887 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-08T23:30:32.889 INFO:teuthology.task.internal:Creating test directory... 2026-03-08T23:30:32.889 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-08T23:30:32.890 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-08T23:30:32.892 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-08T23:30:32.894 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-08T23:30:32.895 INFO:teuthology.task.internal:Creating archive directory... 2026-03-08T23:30:32.895 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-08T23:30:32.937 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-08T23:30:32.941 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-08T23:30:32.942 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-08T23:30:32.942 DEBUG:teuthology.orchestra.run.vm02:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-08T23:30:32.982 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:30:32.982 DEBUG:teuthology.orchestra.run.vm05:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-08T23:30:32.985 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:30:32.985 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-08T23:30:33.025 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-08T23:30:33.032 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-08T23:30:33.035 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-08T23:30:33.036 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-08T23:30:33.039 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-08T23:30:33.040 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-08T23:30:33.041 INFO:teuthology.task.internal:Configuring sudo... 2026-03-08T23:30:33.041 DEBUG:teuthology.orchestra.run.vm02:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-08T23:30:33.081 DEBUG:teuthology.orchestra.run.vm05:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-08T23:30:33.089 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-08T23:30:33.091 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-08T23:30:33.091 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-08T23:30:33.128 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-08T23:30:33.133 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-08T23:30:33.174 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-08T23:30:33.219 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:30:33.219 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-08T23:30:33.272 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-08T23:30:33.275 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-08T23:30:33.322 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:30:33.322 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-08T23:30:33.370 DEBUG:teuthology.orchestra.run.vm02:> sudo service rsyslog restart 2026-03-08T23:30:33.371 DEBUG:teuthology.orchestra.run.vm05:> sudo service rsyslog restart 2026-03-08T23:30:33.425 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-08T23:30:33.427 INFO:teuthology.task.internal:Starting timer... 2026-03-08T23:30:33.427 INFO:teuthology.run_tasks:Running task pcp... 2026-03-08T23:30:33.430 INFO:teuthology.run_tasks:Running task selinux... 2026-03-08T23:30:33.432 INFO:teuthology.task.selinux:Excluding vm02: VMs are not yet supported 2026-03-08T23:30:33.432 INFO:teuthology.task.selinux:Excluding vm05: VMs are not yet supported 2026-03-08T23:30:33.432 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-08T23:30:33.432 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-08T23:30:33.432 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-08T23:30:33.432 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-08T23:30:33.433 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-08T23:30:33.433 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-08T23:30:33.435 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-08T23:30:34.006 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-08T23:30:34.012 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-08T23:30:34.012 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryhbcnrg4w --limit vm02.local,vm05.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-08T23:32:34.584 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm02.local'), Remote(name='ubuntu@vm05.local')] 2026-03-08T23:32:34.584 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm02.local' 2026-03-08T23:32:34.585 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-08T23:32:34.646 DEBUG:teuthology.orchestra.run.vm02:> true 2026-03-08T23:32:34.865 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm02.local' 2026-03-08T23:32:34.865 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm05.local' 2026-03-08T23:32:34.865 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-08T23:32:34.926 DEBUG:teuthology.orchestra.run.vm05:> true 2026-03-08T23:32:35.152 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm05.local' 2026-03-08T23:32:35.153 INFO:teuthology.run_tasks:Running task clock... 2026-03-08T23:32:35.155 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-08T23:32:35.155 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-08T23:32:35.156 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-08T23:32:35.157 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-08T23:32:35.157 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Command line: ntpd -gq 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: ---------------------------------------------------- 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: ntp-4 is maintained by Network Time Foundation, 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: corporation. Support and training for ntp-4 are 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: available at https://www.nwtime.org/support 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: ---------------------------------------------------- 2026-03-08T23:32:35.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: proto: precision = 0.030 usec (-25) 2026-03-08T23:32:35.174 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: basedate set to 2022-02-04 2026-03-08T23:32:35.174 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: gps base set to 2022-02-06 (week 2196) 2026-03-08T23:32:35.174 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-08T23:32:35.174 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-08T23:32:35.174 INFO:teuthology.orchestra.run.vm02.stderr: 8 Mar 23:32:35 ntpd[15991]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 71 days ago 2026-03-08T23:32:35.175 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Listen and drop on 0 v6wildcard [::]:123 2026-03-08T23:32:35.175 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-08T23:32:35.175 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Listen normally on 2 lo 127.0.0.1:123 2026-03-08T23:32:35.175 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Listen normally on 3 ens3 192.168.123.102:123 2026-03-08T23:32:35.175 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Listen normally on 4 lo [::1]:123 2026-03-08T23:32:35.175 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:2%2]:123 2026-03-08T23:32:35.175 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:35 ntpd[15991]: Listening on routing socket on fd #22 for interface updates 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Command line: ntpd -gq 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: ---------------------------------------------------- 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: ntp-4 is maintained by Network Time Foundation, 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: corporation. Support and training for ntp-4 are 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: available at https://www.nwtime.org/support 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: ---------------------------------------------------- 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: proto: precision = 0.029 usec (-25) 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: basedate set to 2022-02-04 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: gps base set to 2022-02-06 (week 2196) 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-08T23:32:35.212 INFO:teuthology.orchestra.run.vm05.stderr: 8 Mar 23:32:35 ntpd[16000]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 71 days ago 2026-03-08T23:32:35.213 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Listen and drop on 0 v6wildcard [::]:123 2026-03-08T23:32:35.213 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-08T23:32:35.213 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Listen normally on 2 lo 127.0.0.1:123 2026-03-08T23:32:35.213 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Listen normally on 3 ens3 192.168.123.105:123 2026-03-08T23:32:35.213 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Listen normally on 4 lo [::1]:123 2026-03-08T23:32:35.214 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:5%2]:123 2026-03-08T23:32:35.214 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:35 ntpd[16000]: Listening on routing socket on fd #22 for interface updates 2026-03-08T23:32:36.174 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:36 ntpd[15991]: Soliciting pool server 88.198.200.96 2026-03-08T23:32:36.212 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:36 ntpd[16000]: Soliciting pool server 88.198.200.96 2026-03-08T23:32:37.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:37 ntpd[15991]: Soliciting pool server 136.243.7.20 2026-03-08T23:32:37.211 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:37 ntpd[16000]: Soliciting pool server 136.243.7.20 2026-03-08T23:32:37.330 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:37 ntpd[15991]: Soliciting pool server 37.114.42.119 2026-03-08T23:32:37.330 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:37 ntpd[16000]: Soliciting pool server 37.114.42.119 2026-03-08T23:32:38.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:38 ntpd[15991]: Soliciting pool server 93.177.65.20 2026-03-08T23:32:38.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:38 ntpd[15991]: Soliciting pool server 141.144.246.224 2026-03-08T23:32:38.173 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:38 ntpd[15991]: Soliciting pool server 144.76.43.40 2026-03-08T23:32:38.211 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:38 ntpd[16000]: Soliciting pool server 93.177.65.20 2026-03-08T23:32:38.211 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:38 ntpd[16000]: Soliciting pool server 141.144.246.224 2026-03-08T23:32:38.211 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:38 ntpd[16000]: Soliciting pool server 144.76.43.40 2026-03-08T23:32:39.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:39 ntpd[15991]: Soliciting pool server 176.9.157.155 2026-03-08T23:32:39.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:39 ntpd[15991]: Soliciting pool server 152.53.184.199 2026-03-08T23:32:39.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:39 ntpd[15991]: Soliciting pool server 162.19.170.154 2026-03-08T23:32:39.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:39 ntpd[15991]: Soliciting pool server 77.90.16.198 2026-03-08T23:32:39.211 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:39 ntpd[16000]: Soliciting pool server 176.9.157.155 2026-03-08T23:32:39.211 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:39 ntpd[16000]: Soliciting pool server 152.53.184.199 2026-03-08T23:32:39.211 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:39 ntpd[16000]: Soliciting pool server 162.19.170.154 2026-03-08T23:32:39.211 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:39 ntpd[16000]: Soliciting pool server 77.90.16.198 2026-03-08T23:32:40.171 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:40 ntpd[15991]: Soliciting pool server 193.175.73.20 2026-03-08T23:32:40.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:40 ntpd[15991]: Soliciting pool server 158.180.28.150 2026-03-08T23:32:40.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:40 ntpd[15991]: Soliciting pool server 185.13.148.71 2026-03-08T23:32:40.172 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:40 ntpd[15991]: Soliciting pool server 91.189.91.157 2026-03-08T23:32:40.210 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:40 ntpd[16000]: Soliciting pool server 193.175.73.20 2026-03-08T23:32:40.210 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:40 ntpd[16000]: Soliciting pool server 158.180.28.150 2026-03-08T23:32:40.210 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:40 ntpd[16000]: Soliciting pool server 185.13.148.71 2026-03-08T23:32:40.210 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:40 ntpd[16000]: Soliciting pool server 91.189.91.157 2026-03-08T23:32:41.171 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:41 ntpd[15991]: Soliciting pool server 185.125.190.57 2026-03-08T23:32:41.171 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:41 ntpd[15991]: Soliciting pool server 80.152.203.134 2026-03-08T23:32:41.171 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:41 ntpd[15991]: Soliciting pool server 139.144.71.56 2026-03-08T23:32:41.209 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:41 ntpd[16000]: Soliciting pool server 185.125.190.57 2026-03-08T23:32:41.209 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:41 ntpd[16000]: Soliciting pool server 80.152.203.134 2026-03-08T23:32:41.210 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:41 ntpd[16000]: Soliciting pool server 139.144.71.56 2026-03-08T23:32:44.190 INFO:teuthology.orchestra.run.vm02.stdout: 8 Mar 23:32:44 ntpd[15991]: ntpd: time slew +0.012267 s 2026-03-08T23:32:44.190 INFO:teuthology.orchestra.run.vm02.stdout:ntpd: time slew +0.012267s 2026-03-08T23:32:44.210 INFO:teuthology.orchestra.run.vm02.stdout: remote refid st t when poll reach delay offset jitter 2026-03-08T23:32:44.210 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================== 2026-03-08T23:32:44.210 INFO:teuthology.orchestra.run.vm02.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:44.210 INFO:teuthology.orchestra.run.vm02.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:44.210 INFO:teuthology.orchestra.run.vm02.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:44.210 INFO:teuthology.orchestra.run.vm02.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:44.210 INFO:teuthology.orchestra.run.vm02.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:46.228 INFO:teuthology.orchestra.run.vm05.stdout: 8 Mar 23:32:46 ntpd[16000]: ntpd: time slew +0.000291 s 2026-03-08T23:32:46.228 INFO:teuthology.orchestra.run.vm05.stdout:ntpd: time slew +0.000291s 2026-03-08T23:32:46.249 INFO:teuthology.orchestra.run.vm05.stdout: remote refid st t when poll reach delay offset jitter 2026-03-08T23:32:46.249 INFO:teuthology.orchestra.run.vm05.stdout:============================================================================== 2026-03-08T23:32:46.249 INFO:teuthology.orchestra.run.vm05.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:46.249 INFO:teuthology.orchestra.run.vm05.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:46.249 INFO:teuthology.orchestra.run.vm05.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:46.249 INFO:teuthology.orchestra.run.vm05.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:46.249 INFO:teuthology.orchestra.run.vm05.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:32:46.249 INFO:teuthology.run_tasks:Running task install... 2026-03-08T23:32:46.251 DEBUG:teuthology.task.install:project ceph 2026-03-08T23:32:46.251 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-08T23:32:46.251 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-08T23:32:46.251 INFO:teuthology.task.install:Using flavor: default 2026-03-08T23:32:46.253 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-08T23:32:46.253 INFO:teuthology.task.install:extra packages: [] 2026-03-08T23:32:46.253 DEBUG:teuthology.orchestra.run.vm02:> sudo apt-key list | grep Ceph 2026-03-08T23:32:46.254 DEBUG:teuthology.orchestra.run.vm05:> sudo apt-key list | grep Ceph 2026-03-08T23:32:46.291 INFO:teuthology.orchestra.run.vm02.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-08T23:32:46.310 INFO:teuthology.orchestra.run.vm02.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-08T23:32:46.310 INFO:teuthology.orchestra.run.vm02.stdout:uid [ unknown] Ceph.com (release key) 2026-03-08T23:32:46.310 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-08T23:32:46.310 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-xmltodict, python3-jmespath on remote deb x86_64 2026-03-08T23:32:46.310 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:32:46.345 INFO:teuthology.orchestra.run.vm05.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-08T23:32:46.349 INFO:teuthology.orchestra.run.vm05.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-08T23:32:46.349 INFO:teuthology.orchestra.run.vm05.stdout:uid [ unknown] Ceph.com (release key) 2026-03-08T23:32:46.349 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-08T23:32:46.349 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-xmltodict, python3-jmespath on remote deb x86_64 2026-03-08T23:32:46.350 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:32:46.940 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/ 2026-03-08T23:32:46.940 INFO:teuthology.task.install.deb:Package version is 19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:32:46.971 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/ 2026-03-08T23:32:46.971 INFO:teuthology.task.install.deb:Package version is 19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:32:47.466 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:32:47.466 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-08T23:32:47.474 DEBUG:teuthology.orchestra.run.vm02:> sudo apt-get update 2026-03-08T23:32:47.508 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:32:47.508 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-08T23:32:47.515 DEBUG:teuthology.orchestra.run.vm05:> sudo apt-get update 2026-03-08T23:32:47.660 INFO:teuthology.orchestra.run.vm02.stdout:Hit:1 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-08T23:32:47.664 INFO:teuthology.orchestra.run.vm02.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-08T23:32:47.672 INFO:teuthology.orchestra.run.vm02.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-08T23:32:47.691 INFO:teuthology.orchestra.run.vm05.stdout:Hit:1 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-08T23:32:47.694 INFO:teuthology.orchestra.run.vm05.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-08T23:32:47.702 INFO:teuthology.orchestra.run.vm05.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-08T23:32:47.779 INFO:teuthology.orchestra.run.vm02.stdout:Hit:4 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-08T23:32:47.798 INFO:teuthology.orchestra.run.vm05.stdout:Hit:4 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-08T23:32:48.096 INFO:teuthology.orchestra.run.vm05.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy InRelease 2026-03-08T23:32:48.105 INFO:teuthology.orchestra.run.vm02.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy InRelease 2026-03-08T23:32:48.209 INFO:teuthology.orchestra.run.vm05.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release [7662 B] 2026-03-08T23:32:48.220 INFO:teuthology.orchestra.run.vm02.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release [7662 B] 2026-03-08T23:32:48.324 INFO:teuthology.orchestra.run.vm05.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-08T23:32:48.337 INFO:teuthology.orchestra.run.vm02.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-08T23:32:48.437 INFO:teuthology.orchestra.run.vm05.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.1 kB] 2026-03-08T23:32:48.453 INFO:teuthology.orchestra.run.vm02.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.1 kB] 2026-03-08T23:32:48.515 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 25.8 kB in 1s (30.4 kB/s) 2026-03-08T23:32:48.530 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 25.8 kB in 1s (28.7 kB/s) 2026-03-08T23:32:49.227 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:32:49.240 DEBUG:teuthology.orchestra.run.vm05:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=19.2.3-678-ge911bdeb-1jammy cephadm=19.2.3-678-ge911bdeb-1jammy ceph-mds=19.2.3-678-ge911bdeb-1jammy ceph-mgr=19.2.3-678-ge911bdeb-1jammy ceph-common=19.2.3-678-ge911bdeb-1jammy ceph-fuse=19.2.3-678-ge911bdeb-1jammy ceph-test=19.2.3-678-ge911bdeb-1jammy ceph-volume=19.2.3-678-ge911bdeb-1jammy radosgw=19.2.3-678-ge911bdeb-1jammy python3-rados=19.2.3-678-ge911bdeb-1jammy python3-rgw=19.2.3-678-ge911bdeb-1jammy python3-cephfs=19.2.3-678-ge911bdeb-1jammy python3-rbd=19.2.3-678-ge911bdeb-1jammy libcephfs2=19.2.3-678-ge911bdeb-1jammy libcephfs-dev=19.2.3-678-ge911bdeb-1jammy librados2=19.2.3-678-ge911bdeb-1jammy librbd1=19.2.3-678-ge911bdeb-1jammy rbd-fuse=19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:32:49.247 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:32:49.260 DEBUG:teuthology.orchestra.run.vm02:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=19.2.3-678-ge911bdeb-1jammy cephadm=19.2.3-678-ge911bdeb-1jammy ceph-mds=19.2.3-678-ge911bdeb-1jammy ceph-mgr=19.2.3-678-ge911bdeb-1jammy ceph-common=19.2.3-678-ge911bdeb-1jammy ceph-fuse=19.2.3-678-ge911bdeb-1jammy ceph-test=19.2.3-678-ge911bdeb-1jammy ceph-volume=19.2.3-678-ge911bdeb-1jammy radosgw=19.2.3-678-ge911bdeb-1jammy python3-rados=19.2.3-678-ge911bdeb-1jammy python3-rgw=19.2.3-678-ge911bdeb-1jammy python3-cephfs=19.2.3-678-ge911bdeb-1jammy python3-rbd=19.2.3-678-ge911bdeb-1jammy libcephfs2=19.2.3-678-ge911bdeb-1jammy libcephfs-dev=19.2.3-678-ge911bdeb-1jammy librados2=19.2.3-678-ge911bdeb-1jammy librbd1=19.2.3-678-ge911bdeb-1jammy rbd-fuse=19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:32:49.274 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:32:49.294 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:32:49.474 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:32:49.474 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:32:49.507 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:32:49.508 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:32:49.662 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:32:49.662 INFO:teuthology.orchestra.run.vm05.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:32:49.663 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-08T23:32:49.663 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:32:49.663 INFO:teuthology.orchestra.run.vm05.stdout:The following additional packages will be installed: 2026-03-08T23:32:49.663 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-08T23:32:49.663 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-08T23:32:49.664 INFO:teuthology.orchestra.run.vm05.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:32:49.664 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-08T23:32:49.664 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsqlite3-mod-ceph 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: libthrift-0.16.0 lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-pytest python3-repoze.lru 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml python3-waitress python3-wcwidth python3-webob 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket python3-webtest python3-werkzeug python3-zc.lockfile 2026-03-08T23:32:49.665 INFO:teuthology.orchestra.run.vm05.stdout: qttranslations5-l10n smartmontools socat unzip xmlstarlet zip 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout:Suggested packages: 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout: python3-influxdb readline-doc python3-beaker python-mako-doc 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout: python-natsort-doc httpd-wsgi libapache2-mod-python libapache2-mod-scgi 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout: libjs-mochikit python-pecan-doc python-psutil-doc subversion 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout: python-pygments-doc ttf-bitstream-vera python-pyinotify-doc python3-dap 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout: python-sklearn-doc ipython3 python-waitress-doc python-webob-doc 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout: python-webtest-doc python-werkzeug-doc python3-watchdog gsmartcontrol 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout: smart-notifier mailx | mailutils 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout:Recommended packages: 2026-03-08T23:32:49.666 INFO:teuthology.orchestra.run.vm05.stdout: btrfs-tools 2026-03-08T23:32:49.708 INFO:teuthology.orchestra.run.vm05.stdout:The following NEW packages will be installed: 2026-03-08T23:32:49.708 INFO:teuthology.orchestra.run.vm05.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-08T23:32:49.708 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-dev libcephfs2 libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: liblua5.3-dev libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 libreadline-dev 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: librgw2 libsqlite3-mod-ceph libthrift-0.16.0 lua-any lua-sec lua-socket 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: lua5.1 luarocks nvme-cli pkg-config python-asyncssh-doc 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-pytest python3-rados python3-rbd 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze.lru python3-requests-oauthlib python3-rgw python3-routes 2026-03-08T23:32:49.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-08T23:32:49.710 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-08T23:32:49.710 INFO:teuthology.orchestra.run.vm05.stdout: python3-threadpoolctl python3-toml python3-waitress python3-wcwidth 2026-03-08T23:32:49.710 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:32:49.710 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse smartmontools 2026-03-08T23:32:49.710 INFO:teuthology.orchestra.run.vm05.stdout: socat unzip xmlstarlet zip 2026-03-08T23:32:49.710 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be upgraded: 2026-03-08T23:32:49.711 INFO:teuthology.orchestra.run.vm05.stdout: librados2 librbd1 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout:The following additional packages will be installed: 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout: liboath0 libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-08T23:32:49.727 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsqlite3-mod-ceph 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: libthrift-0.16.0 lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-pytest python3-repoze.lru 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml python3-waitress python3-wcwidth python3-webob 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket python3-webtest python3-werkzeug python3-zc.lockfile 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: qttranslations5-l10n smartmontools socat unzip xmlstarlet zip 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout:Suggested packages: 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb readline-doc python3-beaker python-mako-doc 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python-natsort-doc httpd-wsgi libapache2-mod-python libapache2-mod-scgi 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: libjs-mochikit python-pecan-doc python-psutil-doc subversion 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python-pygments-doc ttf-bitstream-vera python-pyinotify-doc python3-dap 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python-sklearn-doc ipython3 python-waitress-doc python-webob-doc 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: python-webtest-doc python-werkzeug-doc python3-watchdog gsmartcontrol 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: smart-notifier mailx | mailutils 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout:Recommended packages: 2026-03-08T23:32:49.728 INFO:teuthology.orchestra.run.vm02.stdout: btrfs-tools 2026-03-08T23:32:49.764 INFO:teuthology.orchestra.run.vm02.stdout:The following NEW packages will be installed: 2026-03-08T23:32:49.764 INFO:teuthology.orchestra.run.vm02.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-dev libcephfs2 libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: liblua5.3-dev libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 libreadline-dev 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: librgw2 libsqlite3-mod-ceph libthrift-0.16.0 lua-any lua-sec lua-socket 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: lua5.1 luarocks nvme-cli pkg-config python-asyncssh-doc 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-pytest python3-rados python3-rbd 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze.lru python3-requests-oauthlib python3-rgw python3-routes 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-threadpoolctl python3-toml python3-waitress python3-wcwidth 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse smartmontools 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: socat unzip xmlstarlet zip 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be upgraded: 2026-03-08T23:32:49.765 INFO:teuthology.orchestra.run.vm02.stdout: librados2 librbd1 2026-03-08T23:32:49.959 INFO:teuthology.orchestra.run.vm02.stdout:2 upgraded, 107 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:32:49.959 INFO:teuthology.orchestra.run.vm02.stdout:Need to get 178 MB of archives. 2026-03-08T23:32:49.959 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 782 MB of additional disk space will be used. 2026-03-08T23:32:49.959 INFO:teuthology.orchestra.run.vm02.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-08T23:32:50.115 INFO:teuthology.orchestra.run.vm02.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-08T23:32:50.120 INFO:teuthology.orchestra.run.vm02.stdout:Get:3 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-08T23:32:50.151 INFO:teuthology.orchestra.run.vm02.stdout:Get:4 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-08T23:32:50.206 INFO:teuthology.orchestra.run.vm05.stdout:2 upgraded, 107 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:32:50.206 INFO:teuthology.orchestra.run.vm05.stdout:Need to get 178 MB of archives. 2026-03-08T23:32:50.206 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 782 MB of additional disk space will be used. 2026-03-08T23:32:50.206 INFO:teuthology.orchestra.run.vm05.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-08T23:32:50.245 INFO:teuthology.orchestra.run.vm02.stdout:Get:5 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-08T23:32:50.249 INFO:teuthology.orchestra.run.vm02.stdout:Get:6 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-08T23:32:50.262 INFO:teuthology.orchestra.run.vm02.stdout:Get:7 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-08T23:32:50.265 INFO:teuthology.orchestra.run.vm02.stdout:Get:8 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-08T23:32:50.266 INFO:teuthology.orchestra.run.vm02.stdout:Get:9 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-08T23:32:50.266 INFO:teuthology.orchestra.run.vm02.stdout:Get:10 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-08T23:32:50.267 INFO:teuthology.orchestra.run.vm02.stdout:Get:11 https://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-08T23:32:50.275 INFO:teuthology.orchestra.run.vm02.stdout:Get:12 https://archive.ubuntu.com/ubuntu jammy/main amd64 libreadline-dev amd64 8.1.2-1 [166 kB] 2026-03-08T23:32:50.277 INFO:teuthology.orchestra.run.vm02.stdout:Get:13 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblua5.3-dev amd64 5.3.6-1build1 [167 kB] 2026-03-08T23:32:50.278 INFO:teuthology.orchestra.run.vm02.stdout:Get:14 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua5.1 amd64 5.1.5-8.1build4 [94.6 kB] 2026-03-08T23:32:50.310 INFO:teuthology.orchestra.run.vm02.stdout:Get:15 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-any all 27ubuntu1 [5034 B] 2026-03-08T23:32:50.310 INFO:teuthology.orchestra.run.vm02.stdout:Get:16 https://archive.ubuntu.com/ubuntu jammy/main amd64 zip amd64 3.0-12build2 [176 kB] 2026-03-08T23:32:50.312 INFO:teuthology.orchestra.run.vm02.stdout:Get:17 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 unzip amd64 6.0-26ubuntu3.2 [175 kB] 2026-03-08T23:32:50.313 INFO:teuthology.orchestra.run.vm02.stdout:Get:18 https://archive.ubuntu.com/ubuntu jammy/universe amd64 luarocks all 3.8.0+dfsg1-1 [140 kB] 2026-03-08T23:32:50.314 INFO:teuthology.orchestra.run.vm02.stdout:Get:19 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-08T23:32:50.315 INFO:teuthology.orchestra.run.vm02.stdout:Get:20 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-08T23:32:50.315 INFO:teuthology.orchestra.run.vm02.stdout:Get:21 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-08T23:32:50.316 INFO:teuthology.orchestra.run.vm02.stdout:Get:22 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-08T23:32:50.318 INFO:teuthology.orchestra.run.vm05.stdout:Get:2 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 19.2.3-678-ge911bdeb-1jammy [3257 kB] 2026-03-08T23:32:50.346 INFO:teuthology.orchestra.run.vm02.stdout:Get:23 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-08T23:32:50.347 INFO:teuthology.orchestra.run.vm02.stdout:Get:24 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-08T23:32:50.347 INFO:teuthology.orchestra.run.vm02.stdout:Get:25 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-08T23:32:50.347 INFO:teuthology.orchestra.run.vm02.stdout:Get:26 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-08T23:32:50.366 INFO:teuthology.orchestra.run.vm02.stdout:Get:27 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 19.2.3-678-ge911bdeb-1jammy [3257 kB] 2026-03-08T23:32:50.380 INFO:teuthology.orchestra.run.vm02.stdout:Get:28 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-08T23:32:50.380 INFO:teuthology.orchestra.run.vm02.stdout:Get:29 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-08T23:32:50.382 INFO:teuthology.orchestra.run.vm02.stdout:Get:30 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-08T23:32:50.382 INFO:teuthology.orchestra.run.vm02.stdout:Get:31 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-logutils all 0.3.3-8 [17.6 kB] 2026-03-08T23:32:50.382 INFO:teuthology.orchestra.run.vm02.stdout:Get:32 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-mako all 1.1.3+ds1-2ubuntu0.1 [60.5 kB] 2026-03-08T23:32:50.383 INFO:teuthology.orchestra.run.vm02.stdout:Get:33 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplegeneric all 0.8.1-3 [11.3 kB] 2026-03-08T23:32:50.412 INFO:teuthology.orchestra.run.vm02.stdout:Get:34 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-singledispatch all 3.4.0.3-3 [7320 B] 2026-03-08T23:32:50.413 INFO:teuthology.orchestra.run.vm02.stdout:Get:35 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-08T23:32:50.413 INFO:teuthology.orchestra.run.vm02.stdout:Get:36 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-waitress all 1.4.4-1.1ubuntu1.1 [47.0 kB] 2026-03-08T23:32:50.414 INFO:teuthology.orchestra.run.vm02.stdout:Get:37 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempita all 0.5.2-6ubuntu1 [15.1 kB] 2026-03-08T23:32:50.446 INFO:teuthology.orchestra.run.vm02.stdout:Get:38 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-paste all 3.5.0+dfsg1-1 [456 kB] 2026-03-08T23:32:50.449 INFO:teuthology.orchestra.run.vm02.stdout:Get:39 https://archive.ubuntu.com/ubuntu jammy/main amd64 python-pastedeploy-tpl all 2.1.1-1 [4892 B] 2026-03-08T23:32:50.450 INFO:teuthology.orchestra.run.vm02.stdout:Get:40 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastedeploy all 2.1.1-1 [26.6 kB] 2026-03-08T23:32:50.450 INFO:teuthology.orchestra.run.vm02.stdout:Get:41 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-webtest all 2.0.35-1 [28.5 kB] 2026-03-08T23:32:50.451 INFO:teuthology.orchestra.run.vm02.stdout:Get:42 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pecan all 1.3.3-4ubuntu2 [87.3 kB] 2026-03-08T23:32:50.451 INFO:teuthology.orchestra.run.vm02.stdout:Get:43 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-werkzeug all 2.0.2+dfsg1-1ubuntu0.22.04.3 [181 kB] 2026-03-08T23:32:50.478 INFO:teuthology.orchestra.run.vm02.stdout:Get:44 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-08T23:32:50.479 INFO:teuthology.orchestra.run.vm02.stdout:Get:45 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-08T23:32:50.481 INFO:teuthology.orchestra.run.vm02.stdout:Get:46 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-08T23:32:50.482 INFO:teuthology.orchestra.run.vm02.stdout:Get:47 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-08T23:32:50.511 INFO:teuthology.orchestra.run.vm02.stdout:Get:48 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-08T23:32:50.579 INFO:teuthology.orchestra.run.vm02.stdout:Get:49 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-08T23:32:50.581 INFO:teuthology.orchestra.run.vm02.stdout:Get:50 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-08T23:32:50.581 INFO:teuthology.orchestra.run.vm02.stdout:Get:51 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-08T23:32:50.589 INFO:teuthology.orchestra.run.vm02.stdout:Get:52 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-08T23:32:50.589 INFO:teuthology.orchestra.run.vm02.stdout:Get:53 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-08T23:32:50.590 INFO:teuthology.orchestra.run.vm02.stdout:Get:54 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-08T23:32:50.590 INFO:teuthology.orchestra.run.vm02.stdout:Get:55 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-08T23:32:50.590 INFO:teuthology.orchestra.run.vm02.stdout:Get:56 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-08T23:32:50.591 INFO:teuthology.orchestra.run.vm02.stdout:Get:57 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-08T23:32:50.613 INFO:teuthology.orchestra.run.vm02.stdout:Get:58 https://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-08T23:32:50.626 INFO:teuthology.orchestra.run.vm02.stdout:Get:59 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-08T23:32:50.627 INFO:teuthology.orchestra.run.vm02.stdout:Get:60 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-08T23:32:50.627 INFO:teuthology.orchestra.run.vm02.stdout:Get:61 https://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-08T23:32:50.630 INFO:teuthology.orchestra.run.vm02.stdout:Get:62 https://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-08T23:32:50.632 INFO:teuthology.orchestra.run.vm02.stdout:Get:63 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-socket amd64 3.0~rc1+git+ac3201d-6 [78.9 kB] 2026-03-08T23:32:50.633 INFO:teuthology.orchestra.run.vm02.stdout:Get:64 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-sec amd64 1.0.2-1 [37.6 kB] 2026-03-08T23:32:50.633 INFO:teuthology.orchestra.run.vm02.stdout:Get:65 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-08T23:32:50.646 INFO:teuthology.orchestra.run.vm02.stdout:Get:66 https://archive.ubuntu.com/ubuntu jammy/main amd64 pkg-config amd64 0.29.2-1ubuntu3 [48.2 kB] 2026-03-08T23:32:50.646 INFO:teuthology.orchestra.run.vm02.stdout:Get:67 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-08T23:32:50.671 INFO:teuthology.orchestra.run.vm02.stdout:Get:68 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-08T23:32:50.671 INFO:teuthology.orchestra.run.vm02.stdout:Get:69 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastescript all 2.0.2-4 [54.6 kB] 2026-03-08T23:32:50.672 INFO:teuthology.orchestra.run.vm02.stdout:Get:70 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-08T23:32:50.672 INFO:teuthology.orchestra.run.vm02.stdout:Get:71 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-08T23:32:50.673 INFO:teuthology.orchestra.run.vm02.stdout:Get:72 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-08T23:32:50.674 INFO:teuthology.orchestra.run.vm02.stdout:Get:73 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-08T23:32:50.680 INFO:teuthology.orchestra.run.vm02.stdout:Get:74 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pyinotify all 0.9.6-1.3 [24.8 kB] 2026-03-08T23:32:50.680 INFO:teuthology.orchestra.run.vm02.stdout:Get:75 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-08T23:32:50.713 INFO:teuthology.orchestra.run.vm02.stdout:Get:76 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-08T23:32:50.715 INFO:teuthology.orchestra.run.vm02.stdout:Get:77 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-08T23:32:50.716 INFO:teuthology.orchestra.run.vm02.stdout:Get:78 https://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-08T23:32:50.743 INFO:teuthology.orchestra.run.vm05.stdout:Get:3 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-08T23:32:50.759 INFO:teuthology.orchestra.run.vm05.stdout:Get:4 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-08T23:32:50.781 INFO:teuthology.orchestra.run.vm02.stdout:Get:79 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-08T23:32:50.882 INFO:teuthology.orchestra.run.vm05.stdout:Get:5 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-08T23:32:51.151 INFO:teuthology.orchestra.run.vm05.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 19.2.3-678-ge911bdeb-1jammy [3597 kB] 2026-03-08T23:32:51.180 INFO:teuthology.orchestra.run.vm05.stdout:Get:7 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-08T23:32:51.193 INFO:teuthology.orchestra.run.vm02.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 19.2.3-678-ge911bdeb-1jammy [3597 kB] 2026-03-08T23:32:51.194 INFO:teuthology.orchestra.run.vm05.stdout:Get:8 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-08T23:32:51.230 INFO:teuthology.orchestra.run.vm05.stdout:Get:9 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-08T23:32:51.242 INFO:teuthology.orchestra.run.vm05.stdout:Get:10 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-08T23:32:51.245 INFO:teuthology.orchestra.run.vm05.stdout:Get:11 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-08T23:32:51.245 INFO:teuthology.orchestra.run.vm05.stdout:Get:12 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-08T23:32:51.246 INFO:teuthology.orchestra.run.vm05.stdout:Get:13 https://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-08T23:32:51.276 INFO:teuthology.orchestra.run.vm05.stdout:Get:14 https://archive.ubuntu.com/ubuntu jammy/main amd64 libreadline-dev amd64 8.1.2-1 [166 kB] 2026-03-08T23:32:51.277 INFO:teuthology.orchestra.run.vm05.stdout:Get:15 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 19.2.3-678-ge911bdeb-1jammy [979 kB] 2026-03-08T23:32:51.279 INFO:teuthology.orchestra.run.vm05.stdout:Get:16 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblua5.3-dev amd64 5.3.6-1build1 [167 kB] 2026-03-08T23:32:51.285 INFO:teuthology.orchestra.run.vm05.stdout:Get:17 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua5.1 amd64 5.1.5-8.1build4 [94.6 kB] 2026-03-08T23:32:51.289 INFO:teuthology.orchestra.run.vm05.stdout:Get:18 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 19.2.3-678-ge911bdeb-1jammy [357 kB] 2026-03-08T23:32:51.296 INFO:teuthology.orchestra.run.vm05.stdout:Get:19 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 19.2.3-678-ge911bdeb-1jammy [32.9 kB] 2026-03-08T23:32:51.296 INFO:teuthology.orchestra.run.vm05.stdout:Get:20 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 19.2.3-678-ge911bdeb-1jammy [184 kB] 2026-03-08T23:32:51.299 INFO:teuthology.orchestra.run.vm05.stdout:Get:21 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 19.2.3-678-ge911bdeb-1jammy [70.1 kB] 2026-03-08T23:32:51.300 INFO:teuthology.orchestra.run.vm05.stdout:Get:22 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 19.2.3-678-ge911bdeb-1jammy [334 kB] 2026-03-08T23:32:51.306 INFO:teuthology.orchestra.run.vm05.stdout:Get:23 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 19.2.3-678-ge911bdeb-1jammy [6935 kB] 2026-03-08T23:32:51.317 INFO:teuthology.orchestra.run.vm02.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 19.2.3-678-ge911bdeb-1jammy [979 kB] 2026-03-08T23:32:51.333 INFO:teuthology.orchestra.run.vm02.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 19.2.3-678-ge911bdeb-1jammy [357 kB] 2026-03-08T23:32:51.338 INFO:teuthology.orchestra.run.vm02.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 19.2.3-678-ge911bdeb-1jammy [32.9 kB] 2026-03-08T23:32:51.339 INFO:teuthology.orchestra.run.vm02.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 19.2.3-678-ge911bdeb-1jammy [184 kB] 2026-03-08T23:32:51.342 INFO:teuthology.orchestra.run.vm02.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 19.2.3-678-ge911bdeb-1jammy [70.1 kB] 2026-03-08T23:32:51.343 INFO:teuthology.orchestra.run.vm02.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 19.2.3-678-ge911bdeb-1jammy [334 kB] 2026-03-08T23:32:51.347 INFO:teuthology.orchestra.run.vm02.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 19.2.3-678-ge911bdeb-1jammy [6935 kB] 2026-03-08T23:32:51.383 INFO:teuthology.orchestra.run.vm05.stdout:Get:24 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-any all 27ubuntu1 [5034 B] 2026-03-08T23:32:51.384 INFO:teuthology.orchestra.run.vm05.stdout:Get:25 https://archive.ubuntu.com/ubuntu jammy/main amd64 zip amd64 3.0-12build2 [176 kB] 2026-03-08T23:32:51.386 INFO:teuthology.orchestra.run.vm05.stdout:Get:26 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 unzip amd64 6.0-26ubuntu3.2 [175 kB] 2026-03-08T23:32:51.389 INFO:teuthology.orchestra.run.vm05.stdout:Get:27 https://archive.ubuntu.com/ubuntu jammy/universe amd64 luarocks all 3.8.0+dfsg1-1 [140 kB] 2026-03-08T23:32:51.393 INFO:teuthology.orchestra.run.vm05.stdout:Get:28 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-08T23:32:51.396 INFO:teuthology.orchestra.run.vm05.stdout:Get:29 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-08T23:32:51.396 INFO:teuthology.orchestra.run.vm05.stdout:Get:30 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-08T23:32:51.397 INFO:teuthology.orchestra.run.vm05.stdout:Get:31 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-08T23:32:51.397 INFO:teuthology.orchestra.run.vm05.stdout:Get:32 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-08T23:32:51.499 INFO:teuthology.orchestra.run.vm05.stdout:Get:33 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-08T23:32:51.499 INFO:teuthology.orchestra.run.vm05.stdout:Get:34 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-08T23:32:51.499 INFO:teuthology.orchestra.run.vm05.stdout:Get:35 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-08T23:32:51.499 INFO:teuthology.orchestra.run.vm05.stdout:Get:36 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-08T23:32:51.601 INFO:teuthology.orchestra.run.vm05.stdout:Get:37 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-08T23:32:51.604 INFO:teuthology.orchestra.run.vm05.stdout:Get:38 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-08T23:32:51.604 INFO:teuthology.orchestra.run.vm05.stdout:Get:39 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-logutils all 0.3.3-8 [17.6 kB] 2026-03-08T23:32:51.605 INFO:teuthology.orchestra.run.vm05.stdout:Get:40 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-mako all 1.1.3+ds1-2ubuntu0.1 [60.5 kB] 2026-03-08T23:32:51.605 INFO:teuthology.orchestra.run.vm05.stdout:Get:41 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplegeneric all 0.8.1-3 [11.3 kB] 2026-03-08T23:32:51.606 INFO:teuthology.orchestra.run.vm05.stdout:Get:42 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-singledispatch all 3.4.0.3-3 [7320 B] 2026-03-08T23:32:51.635 INFO:teuthology.orchestra.run.vm05.stdout:Get:43 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 19.2.3-678-ge911bdeb-1jammy [112 kB] 2026-03-08T23:32:51.635 INFO:teuthology.orchestra.run.vm05.stdout:Get:44 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 19.2.3-678-ge911bdeb-1jammy [470 kB] 2026-03-08T23:32:51.638 INFO:teuthology.orchestra.run.vm05.stdout:Get:45 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 19.2.3-678-ge911bdeb-1jammy [26.5 MB] 2026-03-08T23:32:51.669 INFO:teuthology.orchestra.run.vm02.stdout:Get:88 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 19.2.3-678-ge911bdeb-1jammy [112 kB] 2026-03-08T23:32:51.672 INFO:teuthology.orchestra.run.vm02.stdout:Get:89 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 19.2.3-678-ge911bdeb-1jammy [470 kB] 2026-03-08T23:32:51.682 INFO:teuthology.orchestra.run.vm02.stdout:Get:90 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 19.2.3-678-ge911bdeb-1jammy [26.5 MB] 2026-03-08T23:32:51.703 INFO:teuthology.orchestra.run.vm05.stdout:Get:46 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-08T23:32:51.705 INFO:teuthology.orchestra.run.vm05.stdout:Get:47 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-waitress all 1.4.4-1.1ubuntu1.1 [47.0 kB] 2026-03-08T23:32:51.705 INFO:teuthology.orchestra.run.vm05.stdout:Get:48 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempita all 0.5.2-6ubuntu1 [15.1 kB] 2026-03-08T23:32:51.706 INFO:teuthology.orchestra.run.vm05.stdout:Get:49 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-paste all 3.5.0+dfsg1-1 [456 kB] 2026-03-08T23:32:51.806 INFO:teuthology.orchestra.run.vm05.stdout:Get:50 https://archive.ubuntu.com/ubuntu jammy/main amd64 python-pastedeploy-tpl all 2.1.1-1 [4892 B] 2026-03-08T23:32:51.806 INFO:teuthology.orchestra.run.vm05.stdout:Get:51 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastedeploy all 2.1.1-1 [26.6 kB] 2026-03-08T23:32:51.806 INFO:teuthology.orchestra.run.vm05.stdout:Get:52 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-webtest all 2.0.35-1 [28.5 kB] 2026-03-08T23:32:51.807 INFO:teuthology.orchestra.run.vm05.stdout:Get:53 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pecan all 1.3.3-4ubuntu2 [87.3 kB] 2026-03-08T23:32:51.808 INFO:teuthology.orchestra.run.vm05.stdout:Get:54 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-werkzeug all 2.0.2+dfsg1-1ubuntu0.22.04.3 [181 kB] 2026-03-08T23:32:51.811 INFO:teuthology.orchestra.run.vm05.stdout:Get:55 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-08T23:32:51.908 INFO:teuthology.orchestra.run.vm05.stdout:Get:56 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-08T23:32:51.911 INFO:teuthology.orchestra.run.vm05.stdout:Get:57 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-08T23:32:51.911 INFO:teuthology.orchestra.run.vm05.stdout:Get:58 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-08T23:32:51.913 INFO:teuthology.orchestra.run.vm05.stdout:Get:59 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-08T23:32:52.011 INFO:teuthology.orchestra.run.vm05.stdout:Get:60 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-08T23:32:52.014 INFO:teuthology.orchestra.run.vm05.stdout:Get:61 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-08T23:32:52.015 INFO:teuthology.orchestra.run.vm05.stdout:Get:62 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-08T23:32:52.046 INFO:teuthology.orchestra.run.vm05.stdout:Get:63 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-08T23:32:52.047 INFO:teuthology.orchestra.run.vm05.stdout:Get:64 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-08T23:32:52.047 INFO:teuthology.orchestra.run.vm05.stdout:Get:65 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-08T23:32:52.113 INFO:teuthology.orchestra.run.vm05.stdout:Get:66 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-08T23:32:52.114 INFO:teuthology.orchestra.run.vm05.stdout:Get:67 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-08T23:32:52.114 INFO:teuthology.orchestra.run.vm05.stdout:Get:68 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-08T23:32:52.125 INFO:teuthology.orchestra.run.vm05.stdout:Get:69 https://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-08T23:32:52.218 INFO:teuthology.orchestra.run.vm05.stdout:Get:70 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-08T23:32:52.219 INFO:teuthology.orchestra.run.vm05.stdout:Get:71 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-08T23:32:52.220 INFO:teuthology.orchestra.run.vm05.stdout:Get:72 https://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-08T23:32:52.225 INFO:teuthology.orchestra.run.vm05.stdout:Get:73 https://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-08T23:32:52.230 INFO:teuthology.orchestra.run.vm05.stdout:Get:74 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-socket amd64 3.0~rc1+git+ac3201d-6 [78.9 kB] 2026-03-08T23:32:52.230 INFO:teuthology.orchestra.run.vm05.stdout:Get:75 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-sec amd64 1.0.2-1 [37.6 kB] 2026-03-08T23:32:52.321 INFO:teuthology.orchestra.run.vm05.stdout:Get:76 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-08T23:32:52.327 INFO:teuthology.orchestra.run.vm05.stdout:Get:77 https://archive.ubuntu.com/ubuntu jammy/main amd64 pkg-config amd64 0.29.2-1ubuntu3 [48.2 kB] 2026-03-08T23:32:52.327 INFO:teuthology.orchestra.run.vm05.stdout:Get:78 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-08T23:32:52.332 INFO:teuthology.orchestra.run.vm05.stdout:Get:79 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-08T23:32:52.423 INFO:teuthology.orchestra.run.vm05.stdout:Get:80 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastescript all 2.0.2-4 [54.6 kB] 2026-03-08T23:32:52.423 INFO:teuthology.orchestra.run.vm05.stdout:Get:81 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-08T23:32:52.424 INFO:teuthology.orchestra.run.vm05.stdout:Get:82 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-08T23:32:52.427 INFO:teuthology.orchestra.run.vm05.stdout:Get:83 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-08T23:32:52.428 INFO:teuthology.orchestra.run.vm05.stdout:Get:84 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-08T23:32:52.440 INFO:teuthology.orchestra.run.vm05.stdout:Get:85 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pyinotify all 0.9.6-1.3 [24.8 kB] 2026-03-08T23:32:52.527 INFO:teuthology.orchestra.run.vm05.stdout:Get:86 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-08T23:32:52.527 INFO:teuthology.orchestra.run.vm05.stdout:Get:87 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-08T23:32:52.529 INFO:teuthology.orchestra.run.vm05.stdout:Get:88 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-08T23:32:52.531 INFO:teuthology.orchestra.run.vm05.stdout:Get:89 https://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-08T23:32:52.626 INFO:teuthology.orchestra.run.vm05.stdout:Get:90 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 19.2.3-678-ge911bdeb-1jammy [5178 kB] 2026-03-08T23:32:52.628 INFO:teuthology.orchestra.run.vm05.stdout:Get:91 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-08T23:32:52.673 INFO:teuthology.orchestra.run.vm02.stdout:Get:91 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 19.2.3-678-ge911bdeb-1jammy [5178 kB] 2026-03-08T23:32:52.847 INFO:teuthology.orchestra.run.vm05.stdout:Get:92 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 19.2.3-678-ge911bdeb-1jammy [248 kB] 2026-03-08T23:32:52.849 INFO:teuthology.orchestra.run.vm05.stdout:Get:93 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 19.2.3-678-ge911bdeb-1jammy [125 kB] 2026-03-08T23:32:52.850 INFO:teuthology.orchestra.run.vm05.stdout:Get:94 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 19.2.3-678-ge911bdeb-1jammy [1081 kB] 2026-03-08T23:32:52.877 INFO:teuthology.orchestra.run.vm05.stdout:Get:95 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 19.2.3-678-ge911bdeb-1jammy [6239 kB] 2026-03-08T23:32:52.889 INFO:teuthology.orchestra.run.vm02.stdout:Get:92 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 19.2.3-678-ge911bdeb-1jammy [248 kB] 2026-03-08T23:32:52.893 INFO:teuthology.orchestra.run.vm02.stdout:Get:93 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 19.2.3-678-ge911bdeb-1jammy [125 kB] 2026-03-08T23:32:52.895 INFO:teuthology.orchestra.run.vm02.stdout:Get:94 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 19.2.3-678-ge911bdeb-1jammy [1081 kB] 2026-03-08T23:32:52.913 INFO:teuthology.orchestra.run.vm02.stdout:Get:95 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 19.2.3-678-ge911bdeb-1jammy [6239 kB] 2026-03-08T23:32:53.119 INFO:teuthology.orchestra.run.vm05.stdout:Get:96 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 19.2.3-678-ge911bdeb-1jammy [23.0 MB] 2026-03-08T23:32:53.153 INFO:teuthology.orchestra.run.vm02.stdout:Get:96 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 19.2.3-678-ge911bdeb-1jammy [23.0 MB] 2026-03-08T23:32:54.039 INFO:teuthology.orchestra.run.vm05.stdout:Get:97 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 19.2.3-678-ge911bdeb-1jammy [14.2 kB] 2026-03-08T23:32:54.040 INFO:teuthology.orchestra.run.vm05.stdout:Get:98 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 19.2.3-678-ge911bdeb-1jammy [1173 kB] 2026-03-08T23:32:54.057 INFO:teuthology.orchestra.run.vm05.stdout:Get:99 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 19.2.3-678-ge911bdeb-1jammy [2503 kB] 2026-03-08T23:32:54.092 INFO:teuthology.orchestra.run.vm02.stdout:Get:97 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 19.2.3-678-ge911bdeb-1jammy [14.2 kB] 2026-03-08T23:32:54.092 INFO:teuthology.orchestra.run.vm02.stdout:Get:98 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 19.2.3-678-ge911bdeb-1jammy [1173 kB] 2026-03-08T23:32:54.109 INFO:teuthology.orchestra.run.vm02.stdout:Get:99 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 19.2.3-678-ge911bdeb-1jammy [2503 kB] 2026-03-08T23:32:54.172 INFO:teuthology.orchestra.run.vm05.stdout:Get:100 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 19.2.3-678-ge911bdeb-1jammy [798 kB] 2026-03-08T23:32:54.179 INFO:teuthology.orchestra.run.vm05.stdout:Get:101 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 19.2.3-678-ge911bdeb-1jammy [157 kB] 2026-03-08T23:32:54.184 INFO:teuthology.orchestra.run.vm05.stdout:Get:102 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 19.2.3-678-ge911bdeb-1jammy [2396 kB] 2026-03-08T23:32:54.218 INFO:teuthology.orchestra.run.vm02.stdout:Get:100 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 19.2.3-678-ge911bdeb-1jammy [798 kB] 2026-03-08T23:32:54.231 INFO:teuthology.orchestra.run.vm02.stdout:Get:101 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 19.2.3-678-ge911bdeb-1jammy [157 kB] 2026-03-08T23:32:54.233 INFO:teuthology.orchestra.run.vm02.stdout:Get:102 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 19.2.3-678-ge911bdeb-1jammy [2396 kB] 2026-03-08T23:32:54.294 INFO:teuthology.orchestra.run.vm05.stdout:Get:103 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 19.2.3-678-ge911bdeb-1jammy [8625 kB] 2026-03-08T23:32:54.342 INFO:teuthology.orchestra.run.vm02.stdout:Get:103 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 19.2.3-678-ge911bdeb-1jammy [8625 kB] 2026-03-08T23:32:54.632 INFO:teuthology.orchestra.run.vm05.stdout:Get:104 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 19.2.3-678-ge911bdeb-1jammy [14.3 kB] 2026-03-08T23:32:54.632 INFO:teuthology.orchestra.run.vm05.stdout:Get:105 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 19.2.3-678-ge911bdeb-1jammy [52.1 MB] 2026-03-08T23:32:54.835 INFO:teuthology.orchestra.run.vm02.stdout:Get:104 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 19.2.3-678-ge911bdeb-1jammy [14.3 kB] 2026-03-08T23:32:54.835 INFO:teuthology.orchestra.run.vm02.stdout:Get:105 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 19.2.3-678-ge911bdeb-1jammy [52.1 MB] 2026-03-08T23:32:56.619 INFO:teuthology.orchestra.run.vm05.stdout:Get:106 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 19.2.3-678-ge911bdeb-1jammy [135 kB] 2026-03-08T23:32:56.620 INFO:teuthology.orchestra.run.vm05.stdout:Get:107 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 19.2.3-678-ge911bdeb-1jammy [41.0 kB] 2026-03-08T23:32:56.643 INFO:teuthology.orchestra.run.vm05.stdout:Get:108 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 19.2.3-678-ge911bdeb-1jammy [13.7 MB] 2026-03-08T23:32:56.834 INFO:teuthology.orchestra.run.vm02.stdout:Get:106 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 19.2.3-678-ge911bdeb-1jammy [135 kB] 2026-03-08T23:32:56.834 INFO:teuthology.orchestra.run.vm02.stdout:Get:107 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 19.2.3-678-ge911bdeb-1jammy [41.0 kB] 2026-03-08T23:32:56.835 INFO:teuthology.orchestra.run.vm02.stdout:Get:108 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 19.2.3-678-ge911bdeb-1jammy [13.7 MB] 2026-03-08T23:32:57.179 INFO:teuthology.orchestra.run.vm05.stdout:Get:109 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 19.2.3-678-ge911bdeb-1jammy [92.2 kB] 2026-03-08T23:32:57.344 INFO:teuthology.orchestra.run.vm02.stdout:Get:109 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 19.2.3-678-ge911bdeb-1jammy [92.2 kB] 2026-03-08T23:32:57.484 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 178 MB in 7s (23.9 MB/s) 2026-03-08T23:32:57.655 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 178 MB in 8s (23.5 MB/s) 2026-03-08T23:32:58.076 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-08T23:32:58.082 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-08T23:32:58.108 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 111717 files and directories currently installed.) 2026-03-08T23:32:58.110 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../000-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-08T23:32:58.112 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-08T23:32:58.115 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 111717 files and directories currently installed.) 2026-03-08T23:32:58.117 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../000-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-08T23:32:58.120 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-08T23:32:58.133 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-08T23:32:58.138 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../001-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-08T23:32:58.139 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-08T23:32:58.139 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-08T23:32:58.145 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../001-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-08T23:32:58.146 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-08T23:32:58.157 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-08T23:32:58.163 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../002-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-08T23:32:58.163 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-08T23:32:58.164 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-08T23:32:58.171 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../002-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-08T23:32:58.171 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-08T23:32:58.185 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-08T23:32:58.191 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../003-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-08T23:32:58.194 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-08T23:32:58.194 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:32:58.200 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../003-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-08T23:32:58.204 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:32:58.245 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-08T23:32:58.247 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-08T23:32:58.249 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../004-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-08T23:32:58.250 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:32:58.253 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../004-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-08T23:32:58.253 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:32:58.271 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-08T23:32:58.271 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-08T23:32:58.276 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../005-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-08T23:32:58.276 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:32:58.277 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../005-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-08T23:32:58.278 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:32:58.307 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-08T23:32:58.307 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-08T23:32:58.312 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../006-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-08T23:32:58.312 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../006-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-08T23:32:58.313 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-08T23:32:58.313 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-08T23:32:58.339 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../007-librbd1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.341 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../007-librbd1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.342 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librbd1 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-08T23:32:58.343 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librbd1 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-08T23:32:58.446 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../008-librados2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.447 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../008-librados2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.448 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librados2 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-08T23:32:58.450 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librados2 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-08T23:32:58.541 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libnbd0. 2026-03-08T23:32:58.542 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libnbd0. 2026-03-08T23:32:58.546 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../009-libnbd0_1.10.5-1_amd64.deb ... 2026-03-08T23:32:58.547 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-08T23:32:58.547 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../009-libnbd0_1.10.5-1_amd64.deb ... 2026-03-08T23:32:58.548 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-08T23:32:58.562 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs2. 2026-03-08T23:32:58.563 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libcephfs2. 2026-03-08T23:32:58.568 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../010-libcephfs2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.568 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../010-libcephfs2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.568 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.569 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.598 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rados. 2026-03-08T23:32:58.598 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rados. 2026-03-08T23:32:58.604 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../011-python3-rados_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.604 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../011-python3-rados_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.604 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.605 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.625 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-08T23:32:58.625 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-08T23:32:58.630 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../012-python3-ceph-argparse_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:32:58.631 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.631 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../012-python3-ceph-argparse_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:32:58.632 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.644 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cephfs. 2026-03-08T23:32:58.647 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cephfs. 2026-03-08T23:32:58.649 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../013-python3-cephfs_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.650 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.652 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../013-python3-cephfs_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.653 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.666 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-08T23:32:58.669 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-08T23:32:58.671 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../014-python3-ceph-common_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:32:58.672 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.676 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../014-python3-ceph-common_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:32:58.677 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.692 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-08T23:32:58.697 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-08T23:32:58.697 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../015-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-08T23:32:58.698 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-08T23:32:58.702 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../015-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-08T23:32:58.703 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-08T23:32:58.716 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-prettytable. 2026-03-08T23:32:58.721 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../016-python3-prettytable_2.5.0-2_all.deb ... 2026-03-08T23:32:58.722 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-08T23:32:58.722 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-prettytable. 2026-03-08T23:32:58.728 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../016-python3-prettytable_2.5.0-2_all.deb ... 2026-03-08T23:32:58.728 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-08T23:32:58.737 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rbd. 2026-03-08T23:32:58.743 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../017-python3-rbd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.743 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.745 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rbd. 2026-03-08T23:32:58.752 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../017-python3-rbd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.753 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:58.763 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-08T23:32:58.768 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../018-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-08T23:32:58.769 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-08T23:32:58.773 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-08T23:32:58.779 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../018-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-08T23:32:58.780 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-08T23:32:58.790 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libreadline-dev:amd64. 2026-03-08T23:32:58.796 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../019-libreadline-dev_8.1.2-1_amd64.deb ... 2026-03-08T23:32:58.796 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libreadline-dev:amd64 (8.1.2-1) ... 2026-03-08T23:32:58.800 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libreadline-dev:amd64. 2026-03-08T23:32:58.805 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../019-libreadline-dev_8.1.2-1_amd64.deb ... 2026-03-08T23:32:58.806 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libreadline-dev:amd64 (8.1.2-1) ... 2026-03-08T23:32:58.815 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package liblua5.3-dev:amd64. 2026-03-08T23:32:58.820 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../020-liblua5.3-dev_5.3.6-1build1_amd64.deb ... 2026-03-08T23:32:58.820 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-08T23:32:58.824 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package liblua5.3-dev:amd64. 2026-03-08T23:32:58.830 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../020-liblua5.3-dev_5.3.6-1build1_amd64.deb ... 2026-03-08T23:32:58.831 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-08T23:32:58.840 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package lua5.1. 2026-03-08T23:32:58.846 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../021-lua5.1_5.1.5-8.1build4_amd64.deb ... 2026-03-08T23:32:58.846 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking lua5.1 (5.1.5-8.1build4) ... 2026-03-08T23:32:58.850 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package lua5.1. 2026-03-08T23:32:58.856 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../021-lua5.1_5.1.5-8.1build4_amd64.deb ... 2026-03-08T23:32:58.857 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking lua5.1 (5.1.5-8.1build4) ... 2026-03-08T23:32:58.865 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package lua-any. 2026-03-08T23:32:58.870 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../022-lua-any_27ubuntu1_all.deb ... 2026-03-08T23:32:58.871 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking lua-any (27ubuntu1) ... 2026-03-08T23:32:58.876 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package lua-any. 2026-03-08T23:32:58.882 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../022-lua-any_27ubuntu1_all.deb ... 2026-03-08T23:32:58.882 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking lua-any (27ubuntu1) ... 2026-03-08T23:32:58.883 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package zip. 2026-03-08T23:32:58.889 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../023-zip_3.0-12build2_amd64.deb ... 2026-03-08T23:32:58.889 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking zip (3.0-12build2) ... 2026-03-08T23:32:58.896 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package zip. 2026-03-08T23:32:58.901 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../023-zip_3.0-12build2_amd64.deb ... 2026-03-08T23:32:58.902 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking zip (3.0-12build2) ... 2026-03-08T23:32:58.908 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package unzip. 2026-03-08T23:32:58.914 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../024-unzip_6.0-26ubuntu3.2_amd64.deb ... 2026-03-08T23:32:58.915 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking unzip (6.0-26ubuntu3.2) ... 2026-03-08T23:32:58.920 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package unzip. 2026-03-08T23:32:58.926 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../024-unzip_6.0-26ubuntu3.2_amd64.deb ... 2026-03-08T23:32:58.927 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking unzip (6.0-26ubuntu3.2) ... 2026-03-08T23:32:58.934 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package luarocks. 2026-03-08T23:32:58.938 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../025-luarocks_3.8.0+dfsg1-1_all.deb ... 2026-03-08T23:32:58.939 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking luarocks (3.8.0+dfsg1-1) ... 2026-03-08T23:32:58.948 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package luarocks. 2026-03-08T23:32:58.954 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../025-luarocks_3.8.0+dfsg1-1_all.deb ... 2026-03-08T23:32:58.955 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking luarocks (3.8.0+dfsg1-1) ... 2026-03-08T23:32:58.989 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package librgw2. 2026-03-08T23:32:58.994 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../026-librgw2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:58.995 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.003 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package librgw2. 2026-03-08T23:32:59.009 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../026-librgw2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.010 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.161 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rgw. 2026-03-08T23:32:59.161 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rgw. 2026-03-08T23:32:59.166 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../027-python3-rgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.167 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.167 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../027-python3-rgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.168 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.184 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-08T23:32:59.186 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-08T23:32:59.190 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../028-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-08T23:32:59.191 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-08T23:32:59.191 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../028-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-08T23:32:59.192 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-08T23:32:59.206 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libradosstriper1. 2026-03-08T23:32:59.206 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libradosstriper1. 2026-03-08T23:32:59.211 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../029-libradosstriper1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.212 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../029-libradosstriper1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.212 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.212 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.237 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-common. 2026-03-08T23:32:59.238 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-common. 2026-03-08T23:32:59.243 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../030-ceph-common_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.243 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.244 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../030-ceph-common_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.245 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.772 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-base. 2026-03-08T23:32:59.778 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../031-ceph-base_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.783 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-base. 2026-03-08T23:32:59.783 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.785 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../031-ceph-base_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:32:59.789 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:32:59.943 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-08T23:32:59.946 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-08T23:32:59.949 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../032-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-08T23:32:59.949 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-08T23:32:59.952 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../032-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-08T23:32:59.952 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-08T23:32:59.964 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cheroot. 2026-03-08T23:32:59.969 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cheroot. 2026-03-08T23:32:59.970 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../033-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-08T23:32:59.971 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-08T23:32:59.973 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../033-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-08T23:32:59.974 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-08T23:32:59.990 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-08T23:32:59.995 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-08T23:32:59.995 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../034-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-08T23:32:59.996 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-08T23:33:00.000 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../034-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-08T23:33:00.001 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-08T23:33:00.014 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-08T23:33:00.017 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-08T23:33:00.020 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../035-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-08T23:33:00.021 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-08T23:33:00.023 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../035-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-08T23:33:00.025 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-08T23:33:00.036 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-08T23:33:00.041 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-08T23:33:00.042 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../036-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-08T23:33:00.043 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-08T23:33:00.047 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../036-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-08T23:33:00.048 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-08T23:33:00.058 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-tempora. 2026-03-08T23:33:00.063 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../037-python3-tempora_4.1.2-1_all.deb ... 2026-03-08T23:33:00.064 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-08T23:33:00.064 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-tempora. 2026-03-08T23:33:00.070 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../037-python3-tempora_4.1.2-1_all.deb ... 2026-03-08T23:33:00.071 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-08T23:33:00.083 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-portend. 2026-03-08T23:33:00.089 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-portend. 2026-03-08T23:33:00.089 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../038-python3-portend_3.0.0-1_all.deb ... 2026-03-08T23:33:00.090 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-08T23:33:00.094 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../038-python3-portend_3.0.0-1_all.deb ... 2026-03-08T23:33:00.095 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-08T23:33:00.105 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-08T23:33:00.111 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../039-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-08T23:33:00.112 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-08T23:33:00.112 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-08T23:33:00.118 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../039-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-08T23:33:00.119 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-08T23:33:00.128 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-08T23:33:00.133 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../040-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-08T23:33:00.134 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-08T23:33:00.135 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-08T23:33:00.142 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../040-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-08T23:33:00.142 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-08T23:33:00.164 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-natsort. 2026-03-08T23:33:00.169 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../041-python3-natsort_8.0.2-1_all.deb ... 2026-03-08T23:33:00.170 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-08T23:33:00.173 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-natsort. 2026-03-08T23:33:00.179 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../041-python3-natsort_8.0.2-1_all.deb ... 2026-03-08T23:33:00.180 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-08T23:33:00.187 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-logutils. 2026-03-08T23:33:00.193 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../042-python3-logutils_0.3.3-8_all.deb ... 2026-03-08T23:33:00.194 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-logutils (0.3.3-8) ... 2026-03-08T23:33:00.198 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-logutils. 2026-03-08T23:33:00.203 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../042-python3-logutils_0.3.3-8_all.deb ... 2026-03-08T23:33:00.204 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-logutils (0.3.3-8) ... 2026-03-08T23:33:00.209 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-mako. 2026-03-08T23:33:00.214 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../043-python3-mako_1.1.3+ds1-2ubuntu0.1_all.deb ... 2026-03-08T23:33:00.215 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-08T23:33:00.218 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-mako. 2026-03-08T23:33:00.225 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../043-python3-mako_1.1.3+ds1-2ubuntu0.1_all.deb ... 2026-03-08T23:33:00.225 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-08T23:33:00.234 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-simplegeneric. 2026-03-08T23:33:00.240 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../044-python3-simplegeneric_0.8.1-3_all.deb ... 2026-03-08T23:33:00.240 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-simplegeneric (0.8.1-3) ... 2026-03-08T23:33:00.245 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-simplegeneric. 2026-03-08T23:33:00.251 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../044-python3-simplegeneric_0.8.1-3_all.deb ... 2026-03-08T23:33:00.252 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-simplegeneric (0.8.1-3) ... 2026-03-08T23:33:00.256 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-singledispatch. 2026-03-08T23:33:00.262 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../045-python3-singledispatch_3.4.0.3-3_all.deb ... 2026-03-08T23:33:00.263 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-singledispatch (3.4.0.3-3) ... 2026-03-08T23:33:00.267 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-singledispatch. 2026-03-08T23:33:00.273 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../045-python3-singledispatch_3.4.0.3-3_all.deb ... 2026-03-08T23:33:00.273 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-singledispatch (3.4.0.3-3) ... 2026-03-08T23:33:00.278 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-webob. 2026-03-08T23:33:00.284 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../046-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-08T23:33:00.286 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-08T23:33:00.288 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-webob. 2026-03-08T23:33:00.294 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../046-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-08T23:33:00.294 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-08T23:33:00.309 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-waitress. 2026-03-08T23:33:00.314 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-waitress. 2026-03-08T23:33:00.315 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../047-python3-waitress_1.4.4-1.1ubuntu1.1_all.deb ... 2026-03-08T23:33:00.317 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-08T23:33:00.320 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../047-python3-waitress_1.4.4-1.1ubuntu1.1_all.deb ... 2026-03-08T23:33:00.322 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-08T23:33:00.335 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-tempita. 2026-03-08T23:33:00.341 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-tempita. 2026-03-08T23:33:00.342 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../048-python3-tempita_0.5.2-6ubuntu1_all.deb ... 2026-03-08T23:33:00.343 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-tempita (0.5.2-6ubuntu1) ... 2026-03-08T23:33:00.348 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../048-python3-tempita_0.5.2-6ubuntu1_all.deb ... 2026-03-08T23:33:00.348 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-tempita (0.5.2-6ubuntu1) ... 2026-03-08T23:33:00.359 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-paste. 2026-03-08T23:33:00.365 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-paste. 2026-03-08T23:33:00.365 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../049-python3-paste_3.5.0+dfsg1-1_all.deb ... 2026-03-08T23:33:00.366 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-paste (3.5.0+dfsg1-1) ... 2026-03-08T23:33:00.370 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../049-python3-paste_3.5.0+dfsg1-1_all.deb ... 2026-03-08T23:33:00.371 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-paste (3.5.0+dfsg1-1) ... 2026-03-08T23:33:00.405 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python-pastedeploy-tpl. 2026-03-08T23:33:00.407 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python-pastedeploy-tpl. 2026-03-08T23:33:00.410 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../050-python-pastedeploy-tpl_2.1.1-1_all.deb ... 2026-03-08T23:33:00.411 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python-pastedeploy-tpl (2.1.1-1) ... 2026-03-08T23:33:00.412 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../050-python-pastedeploy-tpl_2.1.1-1_all.deb ... 2026-03-08T23:33:00.413 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python-pastedeploy-tpl (2.1.1-1) ... 2026-03-08T23:33:00.426 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pastedeploy. 2026-03-08T23:33:00.429 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pastedeploy. 2026-03-08T23:33:00.431 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../051-python3-pastedeploy_2.1.1-1_all.deb ... 2026-03-08T23:33:00.432 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pastedeploy (2.1.1-1) ... 2026-03-08T23:33:00.434 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../051-python3-pastedeploy_2.1.1-1_all.deb ... 2026-03-08T23:33:00.435 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pastedeploy (2.1.1-1) ... 2026-03-08T23:33:00.449 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-webtest. 2026-03-08T23:33:00.451 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-webtest. 2026-03-08T23:33:00.454 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../052-python3-webtest_2.0.35-1_all.deb ... 2026-03-08T23:33:00.455 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-webtest (2.0.35-1) ... 2026-03-08T23:33:00.458 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../052-python3-webtest_2.0.35-1_all.deb ... 2026-03-08T23:33:00.459 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-webtest (2.0.35-1) ... 2026-03-08T23:33:00.472 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pecan. 2026-03-08T23:33:00.476 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pecan. 2026-03-08T23:33:00.477 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../053-python3-pecan_1.3.3-4ubuntu2_all.deb ... 2026-03-08T23:33:00.478 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pecan (1.3.3-4ubuntu2) ... 2026-03-08T23:33:00.481 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../053-python3-pecan_1.3.3-4ubuntu2_all.deb ... 2026-03-08T23:33:00.482 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pecan (1.3.3-4ubuntu2) ... 2026-03-08T23:33:00.508 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-werkzeug. 2026-03-08T23:33:00.512 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-werkzeug. 2026-03-08T23:33:00.514 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../054-python3-werkzeug_2.0.2+dfsg1-1ubuntu0.22.04.3_all.deb ... 2026-03-08T23:33:00.515 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-08T23:33:00.518 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../054-python3-werkzeug_2.0.2+dfsg1-1ubuntu0.22.04.3_all.deb ... 2026-03-08T23:33:00.519 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-08T23:33:00.540 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-08T23:33:00.543 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-08T23:33:00.546 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../055-ceph-mgr-modules-core_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:00.547 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.548 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../055-ceph-mgr-modules-core_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:00.549 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.591 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-08T23:33:00.593 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-08T23:33:00.596 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../056-libsqlite3-mod-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:00.597 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.599 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../056-libsqlite3-mod-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:00.600 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.614 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr. 2026-03-08T23:33:00.618 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr. 2026-03-08T23:33:00.619 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../057-ceph-mgr_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:00.620 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.624 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../057-ceph-mgr_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:00.625 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.656 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mon. 2026-03-08T23:33:00.657 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mon. 2026-03-08T23:33:00.661 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../058-ceph-mon_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:00.662 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.663 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../058-ceph-mon_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:00.664 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.798 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-08T23:33:00.799 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-08T23:33:00.804 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../059-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-08T23:33:00.805 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../059-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-08T23:33:00.805 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-08T23:33:00.806 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-08T23:33:00.823 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-osd. 2026-03-08T23:33:00.824 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-osd. 2026-03-08T23:33:00.829 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../060-ceph-osd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:00.829 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:00.830 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../060-ceph-osd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:00.831 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.275 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph. 2026-03-08T23:33:01.275 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph. 2026-03-08T23:33:01.281 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../061-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:01.282 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../061-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:01.282 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.283 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.296 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-fuse. 2026-03-08T23:33:01.299 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-fuse. 2026-03-08T23:33:01.302 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../062-ceph-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:01.303 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.305 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../062-ceph-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:01.306 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.339 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mds. 2026-03-08T23:33:01.340 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mds. 2026-03-08T23:33:01.345 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../063-ceph-mds_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:01.346 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.348 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../063-ceph-mds_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:01.348 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.419 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package cephadm. 2026-03-08T23:33:01.419 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package cephadm. 2026-03-08T23:33:01.424 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../064-cephadm_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:01.425 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../064-cephadm_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:01.425 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.425 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.444 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-08T23:33:01.445 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-08T23:33:01.450 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../065-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-08T23:33:01.451 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../065-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-08T23:33:01.451 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-08T23:33:01.451 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-08T23:33:01.477 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-08T23:33:01.482 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-08T23:33:01.482 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../066-ceph-mgr-cephadm_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:01.483 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.486 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../066-ceph-mgr-cephadm_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:01.487 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.508 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-08T23:33:01.512 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-08T23:33:01.514 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../067-python3-repoze.lru_0.7-2_all.deb ... 2026-03-08T23:33:01.515 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-08T23:33:01.517 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../067-python3-repoze.lru_0.7-2_all.deb ... 2026-03-08T23:33:01.518 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-08T23:33:01.530 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-routes. 2026-03-08T23:33:01.533 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-routes. 2026-03-08T23:33:01.536 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../068-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-08T23:33:01.537 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-08T23:33:01.539 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../068-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-08T23:33:01.539 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-08T23:33:01.563 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-08T23:33:01.563 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-08T23:33:01.568 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../069-ceph-mgr-dashboard_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:01.569 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../069-ceph-mgr-dashboard_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:01.569 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:01.570 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:02.086 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-08T23:33:02.087 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-08T23:33:02.090 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../070-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-08T23:33:02.091 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-08T23:33:02.093 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../070-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-08T23:33:02.094 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-08T23:33:02.181 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-joblib. 2026-03-08T23:33:02.183 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-joblib. 2026-03-08T23:33:02.187 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../071-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-08T23:33:02.188 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-08T23:33:02.189 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../071-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-08T23:33:02.190 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-08T23:33:02.226 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-08T23:33:02.229 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-08T23:33:02.231 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../072-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-08T23:33:02.232 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-08T23:33:02.235 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../072-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-08T23:33:02.236 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-08T23:33:02.247 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-sklearn. 2026-03-08T23:33:02.252 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../073-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-08T23:33:02.253 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-08T23:33:02.254 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-sklearn. 2026-03-08T23:33:02.260 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../073-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-08T23:33:02.261 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-08T23:33:02.406 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-08T23:33:02.412 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../074-ceph-mgr-diskprediction-local_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:02.413 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:02.433 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-08T23:33:02.435 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../074-ceph-mgr-diskprediction-local_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:02.436 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:02.845 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cachetools. 2026-03-08T23:33:02.848 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cachetools. 2026-03-08T23:33:02.851 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../075-python3-cachetools_5.0.0-1_all.deb ... 2026-03-08T23:33:02.851 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-08T23:33:02.854 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../075-python3-cachetools_5.0.0-1_all.deb ... 2026-03-08T23:33:02.855 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-08T23:33:02.865 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rsa. 2026-03-08T23:33:02.869 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../076-python3-rsa_4.8-1_all.deb ... 2026-03-08T23:33:02.870 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-08T23:33:02.870 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rsa. 2026-03-08T23:33:02.875 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../076-python3-rsa_4.8-1_all.deb ... 2026-03-08T23:33:02.876 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-08T23:33:02.887 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-google-auth. 2026-03-08T23:33:02.891 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../077-python3-google-auth_1.5.1-3_all.deb ... 2026-03-08T23:33:02.891 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-08T23:33:02.896 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-google-auth. 2026-03-08T23:33:02.903 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../077-python3-google-auth_1.5.1-3_all.deb ... 2026-03-08T23:33:02.904 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-08T23:33:02.909 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-08T23:33:02.912 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../078-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-08T23:33:02.913 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-08T23:33:02.924 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-08T23:33:02.928 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-websocket. 2026-03-08T23:33:02.931 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../078-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-08T23:33:02.932 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-08T23:33:02.932 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../079-python3-websocket_1.2.3-1_all.deb ... 2026-03-08T23:33:02.933 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-08T23:33:02.953 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-websocket. 2026-03-08T23:33:02.953 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-08T23:33:02.959 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../079-python3-websocket_1.2.3-1_all.deb ... 2026-03-08T23:33:02.959 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../080-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-08T23:33:02.960 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-08T23:33:02.974 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-08T23:33:02.982 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-08T23:33:02.988 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../080-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-08T23:33:03.003 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-08T23:33:03.163 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-08T23:33:03.169 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../081-ceph-mgr-k8sevents_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:03.170 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:03.180 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-08T23:33:03.186 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-08T23:33:03.186 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../081-ceph-mgr-k8sevents_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:03.187 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:03.192 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../082-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-08T23:33:03.193 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-08T23:33:03.203 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-08T23:33:03.211 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../082-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-08T23:33:03.212 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-08T23:33:03.213 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-08T23:33:03.220 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../083-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-08T23:33:03.222 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-08T23:33:03.232 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-08T23:33:03.239 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package jq. 2026-03-08T23:33:03.239 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../083-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-08T23:33:03.240 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-08T23:33:03.245 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../084-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-08T23:33:03.246 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-08T23:33:03.256 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package jq. 2026-03-08T23:33:03.263 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package socat. 2026-03-08T23:33:03.263 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../084-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-08T23:33:03.264 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-08T23:33:03.269 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../085-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-08T23:33:03.270 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-08T23:33:03.283 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package socat. 2026-03-08T23:33:03.289 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../085-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-08T23:33:03.289 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-08T23:33:03.296 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package xmlstarlet. 2026-03-08T23:33:03.302 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../086-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-08T23:33:03.303 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-08T23:33:03.314 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package xmlstarlet. 2026-03-08T23:33:03.321 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../086-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-08T23:33:03.323 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-08T23:33:03.358 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-test. 2026-03-08T23:33:03.360 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../087-ceph-test_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:03.361 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:03.375 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-test. 2026-03-08T23:33:03.382 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../087-ceph-test_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:03.383 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:04.661 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-volume. 2026-03-08T23:33:04.663 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-volume. 2026-03-08T23:33:04.667 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../088-ceph-volume_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:04.668 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:04.671 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../088-ceph-volume_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-08T23:33:04.672 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:04.699 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-08T23:33:04.703 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-08T23:33:04.704 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../089-libcephfs-dev_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:04.705 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:04.708 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../089-libcephfs-dev_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:04.709 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:04.721 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package lua-socket:amd64. 2026-03-08T23:33:04.727 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package lua-socket:amd64. 2026-03-08T23:33:04.727 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../090-lua-socket_3.0~rc1+git+ac3201d-6_amd64.deb ... 2026-03-08T23:33:04.728 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-08T23:33:04.733 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../090-lua-socket_3.0~rc1+git+ac3201d-6_amd64.deb ... 2026-03-08T23:33:04.734 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-08T23:33:04.752 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package lua-sec:amd64. 2026-03-08T23:33:04.762 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../091-lua-sec_1.0.2-1_amd64.deb ... 2026-03-08T23:33:04.762 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking lua-sec:amd64 (1.0.2-1) ... 2026-03-08T23:33:04.762 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package lua-sec:amd64. 2026-03-08T23:33:04.764 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../091-lua-sec_1.0.2-1_amd64.deb ... 2026-03-08T23:33:04.765 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking lua-sec:amd64 (1.0.2-1) ... 2026-03-08T23:33:04.780 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package nvme-cli. 2026-03-08T23:33:04.785 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package nvme-cli. 2026-03-08T23:33:04.785 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../092-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-08T23:33:04.786 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-08T23:33:04.791 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../092-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-08T23:33:04.792 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-08T23:33:04.830 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package pkg-config. 2026-03-08T23:33:04.835 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../093-pkg-config_0.29.2-1ubuntu3_amd64.deb ... 2026-03-08T23:33:04.836 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking pkg-config (0.29.2-1ubuntu3) ... 2026-03-08T23:33:04.836 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package pkg-config. 2026-03-08T23:33:04.842 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../093-pkg-config_0.29.2-1ubuntu3_amd64.deb ... 2026-03-08T23:33:04.843 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking pkg-config (0.29.2-1ubuntu3) ... 2026-03-08T23:33:04.854 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-08T23:33:04.859 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-08T23:33:04.861 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../094-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-08T23:33:04.862 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-08T23:33:04.865 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../094-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-08T23:33:04.866 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-08T23:33:04.957 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-08T23:33:04.959 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-08T23:33:04.963 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../095-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-08T23:33:04.964 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-08T23:33:04.964 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../095-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-08T23:33:04.965 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-08T23:33:04.979 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pastescript. 2026-03-08T23:33:04.980 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pastescript. 2026-03-08T23:33:04.983 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../096-python3-pastescript_2.0.2-4_all.deb ... 2026-03-08T23:33:04.984 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pastescript (2.0.2-4) ... 2026-03-08T23:33:04.986 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../096-python3-pastescript_2.0.2-4_all.deb ... 2026-03-08T23:33:04.987 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pastescript (2.0.2-4) ... 2026-03-08T23:33:05.003 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pluggy. 2026-03-08T23:33:05.006 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../097-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-08T23:33:05.007 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-08T23:33:05.009 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pluggy. 2026-03-08T23:33:05.017 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../097-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-08T23:33:05.018 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-08T23:33:05.022 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-psutil. 2026-03-08T23:33:05.026 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../098-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-08T23:33:05.027 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-08T23:33:05.037 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-psutil. 2026-03-08T23:33:05.044 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../098-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-08T23:33:05.045 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-08T23:33:05.051 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-py. 2026-03-08T23:33:05.057 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../099-python3-py_1.10.0-1_all.deb ... 2026-03-08T23:33:05.058 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-08T23:33:05.070 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-py. 2026-03-08T23:33:05.077 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../099-python3-py_1.10.0-1_all.deb ... 2026-03-08T23:33:05.078 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-08T23:33:05.083 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pygments. 2026-03-08T23:33:05.088 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../100-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-08T23:33:05.089 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-08T23:33:05.104 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pygments. 2026-03-08T23:33:05.112 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../100-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-08T23:33:05.113 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-08T23:33:05.162 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pyinotify. 2026-03-08T23:33:05.168 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../101-python3-pyinotify_0.9.6-1.3_all.deb ... 2026-03-08T23:33:05.169 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pyinotify (0.9.6-1.3) ... 2026-03-08T23:33:05.176 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pyinotify. 2026-03-08T23:33:05.182 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../101-python3-pyinotify_0.9.6-1.3_all.deb ... 2026-03-08T23:33:05.183 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pyinotify (0.9.6-1.3) ... 2026-03-08T23:33:05.184 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-toml. 2026-03-08T23:33:05.191 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../102-python3-toml_0.10.2-1_all.deb ... 2026-03-08T23:33:05.192 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-08T23:33:05.200 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-toml. 2026-03-08T23:33:05.208 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../102-python3-toml_0.10.2-1_all.deb ... 2026-03-08T23:33:05.209 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-08T23:33:05.210 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pytest. 2026-03-08T23:33:05.217 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../103-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-08T23:33:05.218 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-08T23:33:05.227 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pytest. 2026-03-08T23:33:05.234 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../103-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-08T23:33:05.236 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-08T23:33:05.253 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-simplejson. 2026-03-08T23:33:05.259 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../104-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-08T23:33:05.260 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-08T23:33:05.267 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-simplejson. 2026-03-08T23:33:05.273 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../104-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-08T23:33:05.274 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-08T23:33:05.283 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-08T23:33:05.288 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../105-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-08T23:33:05.289 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-08T23:33:05.294 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-08T23:33:05.300 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../105-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-08T23:33:05.301 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-08T23:33:05.443 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package radosgw. 2026-03-08T23:33:05.445 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package radosgw. 2026-03-08T23:33:05.450 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../106-radosgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:05.451 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:05.452 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../106-radosgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:05.454 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:05.776 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package rbd-fuse. 2026-03-08T23:33:05.778 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package rbd-fuse. 2026-03-08T23:33:05.780 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../107-rbd-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:05.780 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../107-rbd-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-08T23:33:05.781 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:05.781 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:05.799 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package smartmontools. 2026-03-08T23:33:05.800 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package smartmontools. 2026-03-08T23:33:05.806 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../108-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-08T23:33:05.807 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../108-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-08T23:33:05.814 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-08T23:33:05.816 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-08T23:33:05.862 INFO:teuthology.orchestra.run.vm02.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-08T23:33:05.864 INFO:teuthology.orchestra.run.vm05.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-08T23:33:06.095 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-08T23:33:06.095 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-08T23:33:06.113 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-08T23:33:06.113 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-08T23:33:06.449 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-08T23:33:06.512 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-08T23:33:06.514 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-08T23:33:06.514 INFO:teuthology.orchestra.run.vm02.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-08T23:33:06.576 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-08T23:33:06.580 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-08T23:33:06.583 INFO:teuthology.orchestra.run.vm05.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-08T23:33:06.643 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-08T23:33:06.818 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-08T23:33:06.865 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-08T23:33:07.202 INFO:teuthology.orchestra.run.vm02.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-08T23:33:07.209 INFO:teuthology.orchestra.run.vm02.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-08T23:33:07.210 INFO:teuthology.orchestra.run.vm02.stdout:Setting up cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:07.256 INFO:teuthology.orchestra.run.vm02.stdout:Adding system user cephadm....done 2026-03-08T23:33:07.265 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-08T23:33:07.268 INFO:teuthology.orchestra.run.vm05.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-08T23:33:07.276 INFO:teuthology.orchestra.run.vm05.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-08T23:33:07.277 INFO:teuthology.orchestra.run.vm05.stdout:Setting up cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:07.320 INFO:teuthology.orchestra.run.vm05.stdout:Adding system user cephadm....done 2026-03-08T23:33:07.329 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-08T23:33:07.343 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-08T23:33:07.412 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-08T23:33:07.414 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-08T23:33:07.416 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-08T23:33:07.480 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-08T23:33:07.482 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-08T23:33:07.482 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-08T23:33:07.548 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-08T23:33:07.550 INFO:teuthology.orchestra.run.vm02.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-08T23:33:07.552 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-08T23:33:07.623 INFO:teuthology.orchestra.run.vm05.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-08T23:33:07.625 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-08T23:33:07.646 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-08T23:33:07.715 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-08T23:33:07.776 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-08T23:33:07.834 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-08T23:33:07.843 INFO:teuthology.orchestra.run.vm02.stdout:Setting up unzip (6.0-26ubuntu3.2) ... 2026-03-08T23:33:07.851 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pyinotify (0.9.6-1.3) ... 2026-03-08T23:33:07.906 INFO:teuthology.orchestra.run.vm05.stdout:Setting up unzip (6.0-26ubuntu3.2) ... 2026-03-08T23:33:07.914 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pyinotify (0.9.6-1.3) ... 2026-03-08T23:33:07.922 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-08T23:33:07.984 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-08T23:33:07.990 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:08.050 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:08.060 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-08T23:33:08.062 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-08T23:33:08.064 INFO:teuthology.orchestra.run.vm02.stdout:Setting up lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-08T23:33:08.066 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libreadline-dev:amd64 (8.1.2-1) ... 2026-03-08T23:33:08.068 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-08T23:33:08.070 INFO:teuthology.orchestra.run.vm02.stdout:Setting up lua5.1 (5.1.5-8.1build4) ... 2026-03-08T23:33:08.074 INFO:teuthology.orchestra.run.vm02.stdout:update-alternatives: using /usr/bin/lua5.1 to provide /usr/bin/lua (lua-interpreter) in auto mode 2026-03-08T23:33:08.077 INFO:teuthology.orchestra.run.vm02.stdout:update-alternatives: using /usr/bin/luac5.1 to provide /usr/bin/luac (lua-compiler) in auto mode 2026-03-08T23:33:08.079 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-08T23:33:08.081 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-08T23:33:08.119 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-08T23:33:08.120 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-08T23:33:08.122 INFO:teuthology.orchestra.run.vm05.stdout:Setting up lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-08T23:33:08.124 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libreadline-dev:amd64 (8.1.2-1) ... 2026-03-08T23:33:08.126 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-08T23:33:08.128 INFO:teuthology.orchestra.run.vm05.stdout:Setting up lua5.1 (5.1.5-8.1build4) ... 2026-03-08T23:33:08.133 INFO:teuthology.orchestra.run.vm05.stdout:update-alternatives: using /usr/bin/lua5.1 to provide /usr/bin/lua (lua-interpreter) in auto mode 2026-03-08T23:33:08.135 INFO:teuthology.orchestra.run.vm05.stdout:update-alternatives: using /usr/bin/luac5.1 to provide /usr/bin/luac (lua-compiler) in auto mode 2026-03-08T23:33:08.137 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-08T23:33:08.139 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-08T23:33:08.209 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-08T23:33:08.264 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-08T23:33:08.288 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-08T23:33:08.334 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-08T23:33:08.361 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-08T23:33:08.405 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-08T23:33:08.443 INFO:teuthology.orchestra.run.vm02.stdout:Setting up zip (3.0-12build2) ... 2026-03-08T23:33:08.445 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-08T23:33:08.486 INFO:teuthology.orchestra.run.vm05.stdout:Setting up zip (3.0-12build2) ... 2026-03-08T23:33:08.488 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-08T23:33:08.837 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-tempita (0.5.2-6ubuntu1) ... 2026-03-08T23:33:08.837 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-tempita (0.5.2-6ubuntu1) ... 2026-03-08T23:33:08.911 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python-pastedeploy-tpl (2.1.1-1) ... 2026-03-08T23:33:08.913 INFO:teuthology.orchestra.run.vm02.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-08T23:33:08.918 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python-pastedeploy-tpl (2.1.1-1) ... 2026-03-08T23:33:08.918 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-08T23:33:08.920 INFO:teuthology.orchestra.run.vm05.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-08T23:33:08.922 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-08T23:33:09.013 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-08T23:33:09.013 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-08T23:33:09.188 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-paste (3.5.0+dfsg1-1) ... 2026-03-08T23:33:09.188 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-paste (3.5.0+dfsg1-1) ... 2026-03-08T23:33:09.411 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-08T23:33:09.417 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-08T23:33:09.508 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-08T23:33:09.641 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-08T23:33:09.659 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-08T23:33:09.724 INFO:teuthology.orchestra.run.vm02.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-08T23:33:09.726 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:09.764 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-08T23:33:09.817 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-08T23:33:09.832 INFO:teuthology.orchestra.run.vm05.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-08T23:33:09.834 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:09.919 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-08T23:33:10.382 INFO:teuthology.orchestra.run.vm02.stdout:Setting up pkg-config (0.29.2-1ubuntu3) ... 2026-03-08T23:33:10.405 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:33:10.411 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-08T23:33:10.470 INFO:teuthology.orchestra.run.vm05.stdout:Setting up pkg-config (0.29.2-1ubuntu3) ... 2026-03-08T23:33:10.484 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-08T23:33:10.486 INFO:teuthology.orchestra.run.vm02.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-08T23:33:10.489 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-08T23:33:10.492 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:33:10.497 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-08T23:33:10.560 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-08T23:33:10.576 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-08T23:33:10.578 INFO:teuthology.orchestra.run.vm05.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-08T23:33:10.581 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-08T23:33:10.635 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:33:10.637 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-08T23:33:10.656 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-08T23:33:10.717 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-singledispatch (3.4.0.3-3) ... 2026-03-08T23:33:10.733 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:33:10.744 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-08T23:33:10.792 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-logutils (0.3.3-8) ... 2026-03-08T23:33:10.826 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-singledispatch (3.4.0.3-3) ... 2026-03-08T23:33:10.877 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-08T23:33:10.911 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-logutils (0.3.3-8) ... 2026-03-08T23:33:10.962 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-simplegeneric (0.8.1-3) ... 2026-03-08T23:33:10.997 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-08T23:33:11.041 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-08T23:33:11.080 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-simplegeneric (0.8.1-3) ... 2026-03-08T23:33:11.129 INFO:teuthology.orchestra.run.vm02.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-08T23:33:11.146 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-08T23:33:11.155 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-08T23:33:11.242 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-08T23:33:11.242 INFO:teuthology.orchestra.run.vm05.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-08T23:33:11.249 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-08T23:33:11.250 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-08T23:33:11.326 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-08T23:33:11.342 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-08T23:33:11.352 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-08T23:33:11.424 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-08T23:33:11.433 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-08T23:33:11.532 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-08T23:33:11.532 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-08T23:33:11.606 INFO:teuthology.orchestra.run.vm02.stdout:Setting up liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-08T23:33:11.608 INFO:teuthology.orchestra.run.vm02.stdout:Setting up lua-sec:amd64 (1.0.2-1) ... 2026-03-08T23:33:11.611 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-08T23:33:11.614 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-08T23:33:11.628 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-08T23:33:11.702 INFO:teuthology.orchestra.run.vm05.stdout:Setting up liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-08T23:33:11.704 INFO:teuthology.orchestra.run.vm05.stdout:Setting up lua-sec:amd64 (1.0.2-1) ... 2026-03-08T23:33:11.706 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-08T23:33:11.709 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-08T23:33:11.764 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pastedeploy (2.1.1-1) ... 2026-03-08T23:33:11.843 INFO:teuthology.orchestra.run.vm02.stdout:Setting up lua-any (27ubuntu1) ... 2026-03-08T23:33:11.845 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-08T23:33:11.867 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pastedeploy (2.1.1-1) ... 2026-03-08T23:33:11.920 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:33:11.922 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-08T23:33:11.955 INFO:teuthology.orchestra.run.vm05.stdout:Setting up lua-any (27ubuntu1) ... 2026-03-08T23:33:11.957 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-08T23:33:11.998 INFO:teuthology.orchestra.run.vm02.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-08T23:33:12.000 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-webtest (2.0.35-1) ... 2026-03-08T23:33:12.032 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:33:12.034 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-08T23:33:12.086 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-08T23:33:12.123 INFO:teuthology.orchestra.run.vm05.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-08T23:33:12.125 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-webtest (2.0.35-1) ... 2026-03-08T23:33:12.205 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-08T23:33:12.229 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pastescript (2.0.2-4) ... 2026-03-08T23:33:12.316 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pecan (1.3.3-4ubuntu2) ... 2026-03-08T23:33:12.348 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pastescript (2.0.2-4) ... 2026-03-08T23:33:12.443 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pecan (1.3.3-4ubuntu2) ... 2026-03-08T23:33:12.443 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-08T23:33:12.445 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:12.447 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:12.450 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-08T23:33:12.577 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-08T23:33:12.579 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:12.582 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:12.584 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-08T23:33:13.044 INFO:teuthology.orchestra.run.vm02.stdout:Setting up luarocks (3.8.0+dfsg1-1) ... 2026-03-08T23:33:13.050 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.052 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.055 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.057 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.059 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.117 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-08T23:33:13.120 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-08T23:33:13.181 INFO:teuthology.orchestra.run.vm05.stdout:Setting up luarocks (3.8.0+dfsg1-1) ... 2026-03-08T23:33:13.187 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.190 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.192 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.194 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.196 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.256 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-08T23:33:13.256 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-08T23:33:13.464 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.466 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.468 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.470 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.472 INFO:teuthology.orchestra.run.vm02.stdout:Setting up rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.474 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.476 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.478 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.511 INFO:teuthology.orchestra.run.vm02.stdout:Adding group ceph....done 2026-03-08T23:33:13.546 INFO:teuthology.orchestra.run.vm02.stdout:Adding system user ceph....done 2026-03-08T23:33:13.554 INFO:teuthology.orchestra.run.vm02.stdout:Setting system user ceph properties....done 2026-03-08T23:33:13.558 INFO:teuthology.orchestra.run.vm02.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-08T23:33:13.620 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-08T23:33:13.649 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.652 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.654 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.656 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.658 INFO:teuthology.orchestra.run.vm05.stdout:Setting up rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.661 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.663 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.665 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:13.697 INFO:teuthology.orchestra.run.vm05.stdout:Adding group ceph....done 2026-03-08T23:33:13.736 INFO:teuthology.orchestra.run.vm05.stdout:Adding system user ceph....done 2026-03-08T23:33:13.745 INFO:teuthology.orchestra.run.vm05.stdout:Setting system user ceph properties....done 2026-03-08T23:33:13.749 INFO:teuthology.orchestra.run.vm05.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-08T23:33:13.812 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-08T23:33:13.868 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-08T23:33:14.034 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-08T23:33:14.271 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:14.274 INFO:teuthology.orchestra.run.vm02.stdout:Setting up radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:14.410 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:14.413 INFO:teuthology.orchestra.run.vm05.stdout:Setting up radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:14.523 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-08T23:33:14.524 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-08T23:33:14.676 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-08T23:33:14.676 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-08T23:33:14.898 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:14.978 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-08T23:33:15.069 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:15.150 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-08T23:33:15.322 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:15.383 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-08T23:33:15.383 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-08T23:33:15.516 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:15.581 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-08T23:33:15.582 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-08T23:33:15.715 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:15.773 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-08T23:33:15.773 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-08T23:33:15.960 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.023 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-08T23:33:16.023 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-08T23:33:16.125 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.201 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-08T23:33:16.201 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-08T23:33:16.414 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.490 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-08T23:33:16.490 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-08T23:33:16.552 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.554 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.566 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.624 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-08T23:33:16.625 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-08T23:33:16.905 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.907 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.921 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.962 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.976 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.978 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:16.981 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-08T23:33:16.981 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-08T23:33:16.991 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:17.113 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-08T23:33:17.120 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:33:17.135 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:33:17.212 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for install-info (6.8-4build1) ... 2026-03-08T23:33:17.329 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:17.342 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:17.344 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:17.357 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:33:17.475 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-08T23:33:17.482 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:33:17.496 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:33:17.572 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for install-info (6.8-4build1) ... 2026-03-08T23:33:17.575 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:17.575 INFO:teuthology.orchestra.run.vm02.stdout:Running kernel seems to be up-to-date. 2026-03-08T23:33:17.575 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:17.575 INFO:teuthology.orchestra.run.vm02.stdout:Services to be restarted: 2026-03-08T23:33:17.577 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart packagekit.service 2026-03-08T23:33:17.580 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:17.580 INFO:teuthology.orchestra.run.vm02.stdout:Service restarts being deferred: 2026-03-08T23:33:17.580 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart unattended-upgrades.service 2026-03-08T23:33:17.580 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:17.580 INFO:teuthology.orchestra.run.vm02.stdout:No containers need to be restarted. 2026-03-08T23:33:17.581 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:17.581 INFO:teuthology.orchestra.run.vm02.stdout:No user sessions are running outdated binaries. 2026-03-08T23:33:17.581 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:17.581 INFO:teuthology.orchestra.run.vm02.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-08T23:33:17.925 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:17.925 INFO:teuthology.orchestra.run.vm05.stdout:Running kernel seems to be up-to-date. 2026-03-08T23:33:17.925 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:17.925 INFO:teuthology.orchestra.run.vm05.stdout:Services to be restarted: 2026-03-08T23:33:17.927 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart packagekit.service 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout:Service restarts being deferred: 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart unattended-upgrades.service 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout:No containers need to be restarted. 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout:No user sessions are running outdated binaries. 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:17.931 INFO:teuthology.orchestra.run.vm05.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-08T23:33:18.515 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:33:18.517 DEBUG:teuthology.orchestra.run.vm02:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-xmltodict python3-jmespath 2026-03-08T23:33:18.593 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:33:18.807 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:33:18.807 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:33:18.858 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:33:18.861 DEBUG:teuthology.orchestra.run.vm05:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-xmltodict python3-jmespath 2026-03-08T23:33:18.937 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:33:19.015 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:33:19.016 INFO:teuthology.orchestra.run.vm02.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:33:19.016 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-08T23:33:19.016 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:33:19.033 INFO:teuthology.orchestra.run.vm02.stdout:The following NEW packages will be installed: 2026-03-08T23:33:19.033 INFO:teuthology.orchestra.run.vm02.stdout: python3-jmespath python3-xmltodict 2026-03-08T23:33:19.123 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 2 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:33:19.123 INFO:teuthology.orchestra.run.vm02.stdout:Need to get 34.3 kB of archives. 2026-03-08T23:33:19.123 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 146 kB of additional disk space will be used. 2026-03-08T23:33:19.123 INFO:teuthology.orchestra.run.vm02.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-08T23:33:19.141 INFO:teuthology.orchestra.run.vm02.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-08T23:33:19.144 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:33:19.145 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:33:19.338 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 34.3 kB in 0s (330 kB/s) 2026-03-08T23:33:19.349 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:33:19.350 INFO:teuthology.orchestra.run.vm05.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:33:19.351 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-08T23:33:19.351 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:33:19.355 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jmespath. 2026-03-08T23:33:19.371 INFO:teuthology.orchestra.run.vm05.stdout:The following NEW packages will be installed: 2026-03-08T23:33:19.371 INFO:teuthology.orchestra.run.vm05.stdout: python3-jmespath python3-xmltodict 2026-03-08T23:33:19.384 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118577 files and directories currently installed.) 2026-03-08T23:33:19.386 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-08T23:33:19.387 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-08T23:33:19.403 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-08T23:33:19.409 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-08T23:33:19.409 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-08T23:33:19.436 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-08T23:33:19.503 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-08T23:33:19.821 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 2 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:33:19.821 INFO:teuthology.orchestra.run.vm05.stdout:Need to get 34.3 kB of archives. 2026-03-08T23:33:19.821 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 146 kB of additional disk space will be used. 2026-03-08T23:33:19.821 INFO:teuthology.orchestra.run.vm05.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-08T23:33:19.877 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:19.877 INFO:teuthology.orchestra.run.vm02.stdout:Running kernel seems to be up-to-date. 2026-03-08T23:33:19.877 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:19.877 INFO:teuthology.orchestra.run.vm02.stdout:Services to be restarted: 2026-03-08T23:33:19.881 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart packagekit.service 2026-03-08T23:33:19.884 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:19.884 INFO:teuthology.orchestra.run.vm02.stdout:Service restarts being deferred: 2026-03-08T23:33:19.884 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart unattended-upgrades.service 2026-03-08T23:33:19.885 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:19.885 INFO:teuthology.orchestra.run.vm02.stdout:No containers need to be restarted. 2026-03-08T23:33:19.885 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:19.885 INFO:teuthology.orchestra.run.vm02.stdout:No user sessions are running outdated binaries. 2026-03-08T23:33:19.885 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:33:19.885 INFO:teuthology.orchestra.run.vm02.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-08T23:33:20.035 INFO:teuthology.orchestra.run.vm05.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-08T23:33:20.251 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 34.3 kB in 1s (51.3 kB/s) 2026-03-08T23:33:20.267 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jmespath. 2026-03-08T23:33:20.301 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118577 files and directories currently installed.) 2026-03-08T23:33:20.304 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-08T23:33:20.305 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-08T23:33:20.320 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-08T23:33:20.326 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-08T23:33:20.327 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-08T23:33:20.352 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-08T23:33:20.420 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-08T23:33:20.795 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:20.795 INFO:teuthology.orchestra.run.vm05.stdout:Running kernel seems to be up-to-date. 2026-03-08T23:33:20.795 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:20.795 INFO:teuthology.orchestra.run.vm05.stdout:Services to be restarted: 2026-03-08T23:33:20.798 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart packagekit.service 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout:Service restarts being deferred: 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart unattended-upgrades.service 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout:No containers need to be restarted. 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout:No user sessions are running outdated binaries. 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:33:20.801 INFO:teuthology.orchestra.run.vm05.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-08T23:33:20.903 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:33:20.907 DEBUG:teuthology.parallel:result is None 2026-03-08T23:33:21.654 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:33:21.658 DEBUG:teuthology.parallel:result is None 2026-03-08T23:33:21.659 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:33:22.267 DEBUG:teuthology.orchestra.run.vm02:> dpkg-query -W -f '${Version}' ceph 2026-03-08T23:33:22.276 INFO:teuthology.orchestra.run.vm02.stdout:19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:33:22.277 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:33:22.277 INFO:teuthology.task.install:The correct ceph version 19.2.3-678-ge911bdeb-1jammy is installed. 2026-03-08T23:33:22.278 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:33:22.915 DEBUG:teuthology.orchestra.run.vm05:> dpkg-query -W -f '${Version}' ceph 2026-03-08T23:33:22.924 INFO:teuthology.orchestra.run.vm05.stdout:19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:33:22.924 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678-ge911bdeb-1jammy 2026-03-08T23:33:22.924 INFO:teuthology.task.install:The correct ceph version 19.2.3-678-ge911bdeb-1jammy is installed. 2026-03-08T23:33:22.925 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-08T23:33:22.925 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:33:22.925 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-08T23:33:22.934 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:33:22.934 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-08T23:33:22.974 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-08T23:33:22.974 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:33:22.974 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/daemon-helper 2026-03-08T23:33:22.984 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-08T23:33:23.032 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:33:23.032 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/daemon-helper 2026-03-08T23:33:23.040 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-08T23:33:23.090 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-08T23:33:23.090 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:33:23.090 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-08T23:33:23.097 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-08T23:33:23.148 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:33:23.148 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-08T23:33:23.155 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-08T23:33:23.204 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-08T23:33:23.205 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:33:23.205 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/stdin-killer 2026-03-08T23:33:23.212 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-08T23:33:23.260 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:33:23.260 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/stdin-killer 2026-03-08T23:33:23.268 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-08T23:33:23.317 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-08T23:33:23.360 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 1}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': False}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-08T23:33:23.360 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:33:23.360 INFO:tasks.cephadm:Cluster fsid is 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:33:23.360 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-08T23:33:23.360 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.102', 'mon.b': '192.168.123.105'} 2026-03-08T23:33:23.360 INFO:tasks.cephadm:First mon is mon.a on vm02 2026-03-08T23:33:23.360 INFO:tasks.cephadm:First mgr is a 2026-03-08T23:33:23.360 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-08T23:33:23.360 DEBUG:teuthology.orchestra.run.vm02:> sudo hostname $(hostname -s) 2026-03-08T23:33:23.369 DEBUG:teuthology.orchestra.run.vm05:> sudo hostname $(hostname -s) 2026-03-08T23:33:23.377 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-08T23:33:23.377 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:33:24.007 INFO:tasks.cephadm:builder_project result: [{'url': 'https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/', 'chacra_url': 'https://1.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'ubuntu', 'distro_version': '22.04', 'distro_codename': 'jammy', 'modified': '2026-02-25 19:37:07.680480', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678-ge911bdeb-1jammy', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.98+toko08', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-08T23:33:24.633 INFO:tasks.util.chacra:got chacra host 1.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=ubuntu%2F22.04%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:33:24.634 INFO:tasks.cephadm:Discovered cachra url: https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm 2026-03-08T23:33:24.634 INFO:tasks.cephadm:Downloading cephadm from url: https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm 2026-03-08T23:33:24.634 DEBUG:teuthology.orchestra.run.vm02:> curl --silent -L https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-08T23:33:25.975 INFO:teuthology.orchestra.run.vm02.stdout:-rw-rw-r-- 1 ubuntu ubuntu 795696 Mar 8 23:33 /home/ubuntu/cephtest/cephadm 2026-03-08T23:33:25.975 DEBUG:teuthology.orchestra.run.vm05:> curl --silent -L https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-08T23:33:27.381 INFO:teuthology.orchestra.run.vm05.stdout:-rw-rw-r-- 1 ubuntu ubuntu 795696 Mar 8 23:33 /home/ubuntu/cephtest/cephadm 2026-03-08T23:33:27.381 DEBUG:teuthology.orchestra.run.vm02:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-08T23:33:27.385 DEBUG:teuthology.orchestra.run.vm05:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-08T23:33:27.393 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-08T23:33:27.393 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-08T23:33:27.430 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-08T23:33:27.517 INFO:teuthology.orchestra.run.vm02.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-08T23:33:27.520 INFO:teuthology.orchestra.run.vm05.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-08T23:34:10.637 INFO:teuthology.orchestra.run.vm02.stdout:{ 2026-03-08T23:34:10.637 INFO:teuthology.orchestra.run.vm02.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-08T23:34:10.637 INFO:teuthology.orchestra.run.vm02.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-08T23:34:10.637 INFO:teuthology.orchestra.run.vm02.stdout: "repo_digests": [ 2026-03-08T23:34:10.637 INFO:teuthology.orchestra.run.vm02.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-08T23:34:10.637 INFO:teuthology.orchestra.run.vm02.stdout: ] 2026-03-08T23:34:10.637 INFO:teuthology.orchestra.run.vm02.stdout:} 2026-03-08T23:34:10.778 INFO:teuthology.orchestra.run.vm05.stdout:{ 2026-03-08T23:34:10.778 INFO:teuthology.orchestra.run.vm05.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-08T23:34:10.779 INFO:teuthology.orchestra.run.vm05.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-08T23:34:10.779 INFO:teuthology.orchestra.run.vm05.stdout: "repo_digests": [ 2026-03-08T23:34:10.779 INFO:teuthology.orchestra.run.vm05.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-08T23:34:10.779 INFO:teuthology.orchestra.run.vm05.stdout: ] 2026-03-08T23:34:10.779 INFO:teuthology.orchestra.run.vm05.stdout:} 2026-03-08T23:34:10.794 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph 2026-03-08T23:34:10.801 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /etc/ceph 2026-03-08T23:34:10.809 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 777 /etc/ceph 2026-03-08T23:34:10.853 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 777 /etc/ceph 2026-03-08T23:34:10.860 INFO:tasks.cephadm:Writing seed config... 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [global] mon election default strategy = 1 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = False 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-08T23:34:10.861 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-08T23:34:10.862 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:34:10.862 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-08T23:34:10.900 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 32c51f16-1b47-11f1-bdef-c3bc34602109 mon election default strategy = 1 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = False [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-08T23:34:10.900 DEBUG:teuthology.orchestra.run.vm02:mon.a> sudo journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service 2026-03-08T23:34:10.942 DEBUG:teuthology.orchestra.run.vm02:mgr.a> sudo journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.a.service 2026-03-08T23:34:10.985 INFO:tasks.cephadm:Bootstrapping... 2026-03-08T23:34:10.986 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.102 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-08T23:34:11.122 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-08T23:34:11.122 INFO:teuthology.orchestra.run.vm02.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '32c51f16-1b47-11f1-bdef-c3bc34602109', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.102', '--skip-admin-label'] 2026-03-08T23:34:11.122 INFO:teuthology.orchestra.run.vm02.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-08T23:34:11.123 INFO:teuthology.orchestra.run.vm02.stdout:Verifying podman|docker is present... 2026-03-08T23:34:11.123 INFO:teuthology.orchestra.run.vm02.stdout:Verifying lvm2 is present... 2026-03-08T23:34:11.123 INFO:teuthology.orchestra.run.vm02.stdout:Verifying time synchronization is in place... 2026-03-08T23:34:11.126 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-08T23:34:11.126 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-08T23:34:11.128 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-08T23:34:11.128 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-08T23:34:11.131 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-08T23:34:11.131 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-08T23:34:11.134 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-08T23:34:11.134 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-08T23:34:11.136 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-08T23:34:11.137 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout masked 2026-03-08T23:34:11.139 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-08T23:34:11.139 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-08T23:34:11.141 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-08T23:34:11.141 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-08T23:34:11.143 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-08T23:34:11.144 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-08T23:34:11.147 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-08T23:34:11.150 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-08T23:34:11.150 INFO:teuthology.orchestra.run.vm02.stdout:Unit ntp.service is enabled and running 2026-03-08T23:34:11.150 INFO:teuthology.orchestra.run.vm02.stdout:Repeating the final host check... 2026-03-08T23:34:11.150 INFO:teuthology.orchestra.run.vm02.stdout:docker (/usr/bin/docker) is present 2026-03-08T23:34:11.150 INFO:teuthology.orchestra.run.vm02.stdout:systemctl is present 2026-03-08T23:34:11.150 INFO:teuthology.orchestra.run.vm02.stdout:lvcreate is present 2026-03-08T23:34:11.152 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-08T23:34:11.152 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-08T23:34:11.155 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-08T23:34:11.155 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-08T23:34:11.158 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-08T23:34:11.158 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-08T23:34:11.161 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-08T23:34:11.161 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-08T23:34:11.164 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-08T23:34:11.164 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout masked 2026-03-08T23:34:11.167 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-08T23:34:11.167 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-08T23:34:11.170 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-08T23:34:11.170 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-08T23:34:11.172 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-08T23:34:11.172 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-08T23:34:11.175 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:Unit ntp.service is enabled and running 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:Host looks OK 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:Cluster fsid: 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:Acquiring lock 139860265208368 on /run/cephadm/32c51f16-1b47-11f1-bdef-c3bc34602109.lock 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:Lock 139860265208368 acquired on /run/cephadm/32c51f16-1b47-11f1-bdef-c3bc34602109.lock 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 3300 ... 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 6789 ... 2026-03-08T23:34:11.179 INFO:teuthology.orchestra.run.vm02.stdout:Base mon IP(s) is [192.168.123.102:3300, 192.168.123.102:6789], mon addrv is [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-08T23:34:11.180 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.102 metric 100 2026-03-08T23:34:11.180 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-08T23:34:11.180 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.102 metric 100 2026-03-08T23:34:11.180 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.102 metric 100 2026-03-08T23:34:11.181 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-08T23:34:11.181 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout fe80::/64 dev ens3 proto kernel metric 256 pref medium 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 2: ens3: mtu 1500 state UP qlen 1000 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout inet6 fe80::5055:ff:fe00:2/64 scope link 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.1/32` 2026-03-08T23:34:11.183 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.1/32` 2026-03-08T23:34:11.184 INFO:teuthology.orchestra.run.vm02.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24', '192.168.123.1/32', '192.168.123.1/32'] 2026-03-08T23:34:11.184 INFO:teuthology.orchestra.run.vm02.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-08T23:34:11.184 INFO:teuthology.orchestra.run.vm02.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-08T23:34:12.152 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout e911bdebe5c8faa3800735d1568fcdca65db60df: Pulling from ceph-ci/ceph 2026-03-08T23:34:12.152 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout Digest: sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-08T23:34:12.152 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout Status: Image is up to date for quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:34:12.152 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-08T23:34:12.314 INFO:teuthology.orchestra.run.vm02.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-08T23:34:12.314 INFO:teuthology.orchestra.run.vm02.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-08T23:34:12.314 INFO:teuthology.orchestra.run.vm02.stdout:Extracting ceph user uid/gid from container image... 2026-03-08T23:34:12.461 INFO:teuthology.orchestra.run.vm02.stdout:stat: stdout 167 167 2026-03-08T23:34:12.461 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial keys... 2026-03-08T23:34:12.564 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQB0B65pxMxHIBAAAWwQwDK//fKibI48ZGhFbA== 2026-03-08T23:34:12.684 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQB0B65pZP/mJRAAGdWxSUajeXkIVM7kypFP0g== 2026-03-08T23:34:12.788 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQB0B65pd548LRAAacN49kDn4C1vFw+J08dv0g== 2026-03-08T23:34:12.788 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial monmap... 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool for a [v2:192.168.123.102:3300,v1:192.168.123.102:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:setting min_mon_release = quincy 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: set fsid to 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:12.888 INFO:teuthology.orchestra.run.vm02.stdout:Creating mon... 2026-03-08T23:34:13.005 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.960+0000 7f5875937d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-08T23:34:13.005 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.960+0000 7f5875937d80 1 imported monmap: 2026-03-08T23:34:13.005 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr epoch 0 2026-03-08T23:34:13.005 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:13.005 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr last_changed 2026-03-08T23:34:12.859532+0000 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr min_mon_release 17 (quincy) 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr election_strategy: 1 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.960+0000 7f5875937d80 0 /usr/bin/ceph-mon: set fsid to 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Git sha 0 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: DB SUMMARY 2026-03-08T23:34:13.009 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: DB Session ID: 5XFUN5O8GFDL8DTAPB8K 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 0, files: 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.error_if_exists: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.create_if_missing: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.env: 0x55bcac16adc0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.info_log: 0x55bcb8e7cda0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.statistics: (nil) 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.use_fsync: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.db_log_dir: 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.wal_dir: 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.write_buffer_manager: 0x55bcb8e735e0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.unordered_write: 0 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-08T23:34:13.010 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.row_cache: None 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.wal_filter: None 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.two_write_queues: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.wal_compression: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.atomic_flush: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_open_files: -1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Compression algorithms supported: 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: kZSTD supported: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: kXpressCompression supported: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: kZlibCompression supported: 1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: [db/db_impl/db_impl_open.cc:317] Creating manifest 1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.merge_operator: 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_filter: None 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55bcb8e6f520) 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks: 1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks_with_high_priority: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr pin_top_level_index_and_filter: 1 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_type: 0 2026-03-08T23:34:13.011 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr data_block_index_type: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_shortening: 1 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr data_block_hash_table_util_ratio: 0.750000 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr checksum: 4 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr no_block_cache: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache: 0x55bcb8e95350 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_name: BinnedLRUCache 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_options: 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr capacity : 536870912 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr num_shard_bits : 4 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr strict_capacity_limit : 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr high_pri_pool_ratio: 0.000 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_compressed: (nil) 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr persistent_cache: (nil) 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_size: 4096 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_size_deviation: 10 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_restart_interval: 16 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_block_restart_interval: 1 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr metadata_block_size: 4096 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr partition_filters: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr use_delta_encoding: 1 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr filter_policy: bloomfilter 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr whole_key_filtering: 1 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr verify_compression: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr read_amp_bytes_per_bit: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr format_version: 5 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr enable_index_compression: 1 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_align: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr max_auto_readahead_size: 262144 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr prepopulate_block_cache: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr initial_auto_readahead_size: 8192 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr num_file_reads_for_auto_readahead: 2 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression: NoCompression 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.num_levels: 7 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-08T23:34:13.012 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.bloom_locality: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.ttl: 2592000 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.enable_blob_files: false 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.min_blob_size: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.964+0000 7f5875937d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.968+0000 7f5875937d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 succeeded,manifest_file_number is 1, next_file_number is 3, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.968+0000 7f5875937d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.968+0000 7f5875937d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: bf9a7919-2cb0-4dde-8761-0dace08ae027 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.968+0000 7f5875937d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 5 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.968+0000 7f5875937d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55bcb8e96e00 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.968+0000 7f5875937d80 4 rocksdb: DB pointer 0x55bcb8f7a000 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.972+0000 7f586d0c1640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-08T23:34:13.013 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.972+0000 7f586d0c1640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** DB Stats ** 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Flush(GB): cumulative 0.000, interval 0.000 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(GB): cumulative 0.000, interval 0.000 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(Total Files): cumulative 0, interval 0 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(L0 Files): cumulative 0, interval 0 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(Keys): cumulative 0, interval 0 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Block cache BinnedLRUCache@0x55bcb8e95350#7 capacity: 512.00 MB usage: 0.00 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1e-05 secs_since: 0 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Block cache entry stats(count,size,portion): Misc(1,0.00 KB,0%) 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** File Read Latency Histogram By Level [default] ** 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.972+0000 7f5875937d80 4 rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.972+0000 7f5875937d80 4 rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-08T23:34:12.972+0000 7f5875937d80 0 /usr/bin/ceph-mon: created monfs at /var/lib/ceph/mon/ceph-a for mon.a 2026-03-08T23:34:13.014 INFO:teuthology.orchestra.run.vm02.stdout:create mon.a on 2026-03-08T23:34:13.188 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Removed /etc/systemd/system/multi-user.target.wants/ceph.target. 2026-03-08T23:34:13.353 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-08T23:34:13.536 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109.target → /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109.target. 2026-03-08T23:34:13.536 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109.target → /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109.target. 2026-03-08T23:34:13.707 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a 2026-03-08T23:34:13.707 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service: Unit ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service not loaded. 2026-03-08T23:34:13.872 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109.target.wants/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service → /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service. 2026-03-08T23:34:13.881 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-08T23:34:13.881 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-08T23:34:13.881 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon to start... 2026-03-08T23:34:13.881 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon... 2026-03-08T23:34:14.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:14 vm02 bash[20133]: cluster 2026-03-08T23:34:14.075303+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout cluster: 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout id: 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout services: 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.231867s) 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout data: 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pgs: 2026-03-08T23:34:14.370 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:14.371 INFO:teuthology.orchestra.run.vm02.stdout:mon is available 2026-03-08T23:34:14.371 INFO:teuthology.orchestra.run.vm02.stdout:Assimilating anything we can from ceph.conf... 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = False 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-08T23:34:14.597 INFO:teuthology.orchestra.run.vm02.stdout:Generating new minimal ceph.conf... 2026-03-08T23:34:14.770 INFO:teuthology.orchestra.run.vm02.stdout:Restarting the monitor... 2026-03-08T23:34:14.946 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:14 vm02 systemd[1]: Stopping Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:34:14.946 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:14 vm02 bash[20133]: debug 2026-03-08T23:34:14.808+0000 7f2d8fbce640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-08T23:34:14.946 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:14 vm02 bash[20133]: debug 2026-03-08T23:34:14.808+0000 7f2d8fbce640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-08T23:34:15.018 INFO:teuthology.orchestra.run.vm02.stdout:Setting public_network to 192.168.123.1/32,192.168.123.0/24 in mon config section 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:14 vm02 bash[20526]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mon-a 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 systemd[1]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service: Deactivated successfully. 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 systemd[1]: Stopped Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 systemd[1]: Started Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.164+0000 7f3198192d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.164+0000 7f3198192d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.164+0000 7f3198192d80 0 pidfile_write: ignore empty --pid-file 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.164+0000 7f3198192d80 0 load: jerasure load: lrc 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Git sha 0 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: DB SUMMARY 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: DB Session ID: 8BNYGBWR6USGOL7HV5B8 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: CURRENT file: CURRENT 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75507 ; 2026-03-08T23:34:15.243 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.error_if_exists: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.create_if_missing: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.env: 0x55dce5f99dc0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.info_log: 0x55dd0042a700 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.statistics: (nil) 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.use_fsync: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.db_log_dir: 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.wal_dir: 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.write_buffer_manager: 0x55dd0042f900 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.unordered_write: 0 2026-03-08T23:34:15.244 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.row_cache: None 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.wal_filter: None 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.two_write_queues: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.wal_compression: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.atomic_flush: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_open_files: -1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Compression algorithms supported: 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: kZSTD supported: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: kXpressCompression supported: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: kZlibCompression supported: 1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-08T23:34:15.245 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.merge_operator: 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_filter: None 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55dd0042a640) 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cache_index_and_filter_blocks: 1 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: pin_top_level_index_and_filter: 1 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: index_type: 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: data_block_index_type: 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: index_shortening: 1 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: data_block_hash_table_util_ratio: 0.750000 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: checksum: 4 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: no_block_cache: 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: block_cache: 0x55dd00451350 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: block_cache_name: BinnedLRUCache 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: block_cache_options: 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: capacity : 536870912 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: num_shard_bits : 4 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: strict_capacity_limit : 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: high_pri_pool_ratio: 0.000 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: block_cache_compressed: (nil) 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: persistent_cache: (nil) 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: block_size: 4096 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: block_size_deviation: 10 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: block_restart_interval: 16 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: index_block_restart_interval: 1 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: metadata_block_size: 4096 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: partition_filters: 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: use_delta_encoding: 1 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: filter_policy: bloomfilter 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: whole_key_filtering: 1 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: verify_compression: 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: read_amp_bytes_per_bit: 0 2026-03-08T23:34:15.246 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: format_version: 5 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: enable_index_compression: 1 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: block_align: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: max_auto_readahead_size: 262144 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: prepopulate_block_cache: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: initial_auto_readahead_size: 8192 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: num_file_reads_for_auto_readahead: 2 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression: NoCompression 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.num_levels: 7 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-08T23:34:15.247 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.bloom_locality: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-08T23:34:15.248 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.ttl: 2592000 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.enable_blob_files: false 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.min_blob_size: 0 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.168+0000 7f3198192d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.172+0000 7f3198192d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.172+0000 7f3198192d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.172+0000 7f3198192d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: bf9a7919-2cb0-4dde-8761-0dace08ae027 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.172+0000 7f3198192d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773012855175982, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.172+0000 7f3198192d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.172+0000 7f3198192d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773012855177878, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72588, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 70867, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9705, "raw_average_key_size": 49, "raw_value_size": 65346, "raw_average_value_size": 333, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773012855, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "bf9a7919-2cb0-4dde-8761-0dace08ae027", "db_session_id": "8BNYGBWR6USGOL7HV5B8", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.172+0000 7f3198192d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773012855177948, "job": 1, "event": "recovery_finished"} 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.172+0000 7f3198192d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.180+0000 7f3198192d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.180+0000 7f3198192d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55dd00452e00 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.180+0000 7f3198192d80 4 rocksdb: DB pointer 0x55dd00568000 2026-03-08T23:34:15.249 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: debug 2026-03-08T23:34:15.180+0000 7f3198192d80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] at bind addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:15.472 INFO:teuthology.orchestra.run.vm02.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-08T23:34:15.473 INFO:teuthology.orchestra.run.vm02.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-08T23:34:15.473 INFO:teuthology.orchestra.run.vm02.stdout:Creating mgr... 2026-03-08T23:34:15.473 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-08T23:34:15.473 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-08T23:34:15.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191286+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191286+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191342+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191342+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191349+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191349+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191354+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-08T23:34:12.859532+0000 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191354+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-08T23:34:12.859532+0000 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191363+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191363+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191368+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191368+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191374+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191374+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191379+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191379+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191699+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191699+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191718+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.191718+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.192321+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-08T23:34:15.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 bash[20612]: cluster 2026-03-08T23:34:15.192321+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-08T23:34:15.680 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.a 2026-03-08T23:34:15.680 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.a.service: Unit ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.a.service not loaded. 2026-03-08T23:34:15.848 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109.target.wants/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.a.service → /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service. 2026-03-08T23:34:15.859 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-08T23:34:15.859 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-08T23:34:15.859 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-08T23:34:15.859 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-08T23:34:15.859 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr to start... 2026-03-08T23:34:15.859 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr... 2026-03-08T23:34:15.878 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:34:15.878 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:15 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "32c51f16-1b47-11f1-bdef-c3bc34602109", 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-08T23:34:16.084 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-08T23:34:14:089156+0000", 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-08T23:34:14.090156+0000", 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-08T23:34:16.085 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (1/15)... 2026-03-08T23:34:16.132 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20891]: debug 2026-03-08T23:34:16.092+0000 7f4d89551140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-08T23:34:16.435 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20891]: debug 2026-03-08T23:34:16.128+0000 7f4d89551140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-08T23:34:16.435 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20891]: debug 2026-03-08T23:34:16.260+0000 7f4d89551140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-08T23:34:16.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20891]: debug 2026-03-08T23:34:16.544+0000 7f4d89551140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-08T23:34:16.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20612]: audit 2026-03-08T23:34:15.436282+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1454493050' entity='client.admin' 2026-03-08T23:34:16.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20612]: audit 2026-03-08T23:34:15.436282+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1454493050' entity='client.admin' 2026-03-08T23:34:16.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20612]: audit 2026-03-08T23:34:16.038600+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4188746370' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:34:16.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20612]: audit 2026-03-08T23:34:16.038600+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4188746370' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:34:17.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:16 vm02 bash[20891]: debug 2026-03-08T23:34:16.980+0000 7f4d89551140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-08T23:34:17.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.060+0000 7f4d89551140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-08T23:34:17.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-08T23:34:17.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-08T23:34:17.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: from numpy import show_config as show_numpy_config 2026-03-08T23:34:17.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.184+0000 7f4d89551140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-08T23:34:17.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.332+0000 7f4d89551140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-08T23:34:17.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.372+0000 7f4d89551140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-08T23:34:17.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.408+0000 7f4d89551140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-08T23:34:17.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.448+0000 7f4d89551140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-08T23:34:17.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.496+0000 7f4d89551140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-08T23:34:18.222 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.928+0000 7f4d89551140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-08T23:34:18.222 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:17 vm02 bash[20891]: debug 2026-03-08T23:34:17.964+0000 7f4d89551140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-08T23:34:18.222 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.000+0000 7f4d89551140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-08T23:34:18.222 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.172+0000 7f4d89551140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-08T23:34:18.324 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:18.324 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "32c51f16-1b47-11f1-bdef-c3bc34602109", 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-08T23:34:14:089156+0000", 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-08T23:34:14.090156+0000", 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-08T23:34:18.325 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-08T23:34:18.326 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (2/15)... 2026-03-08T23:34:18.542 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20612]: audit 2026-03-08T23:34:18.278146+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/2262746306' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:34:18.542 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20612]: audit 2026-03-08T23:34:18.278146+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/2262746306' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:34:18.542 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.220+0000 7f4d89551140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-08T23:34:18.542 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.260+0000 7f4d89551140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-08T23:34:18.542 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.380+0000 7f4d89551140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-08T23:34:18.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.540+0000 7f4d89551140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-08T23:34:18.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.712+0000 7f4d89551140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-08T23:34:18.798 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.748+0000 7f4d89551140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-08T23:34:18.798 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.788+0000 7f4d89551140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-08T23:34:19.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:18 vm02 bash[20891]: debug 2026-03-08T23:34:18.936+0000 7f4d89551140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-08T23:34:19.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20891]: debug 2026-03-08T23:34:19.168+0000 7f4d89551140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: cluster 2026-03-08T23:34:19.173138+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: cluster 2026-03-08T23:34:19.173138+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: cluster 2026-03-08T23:34:19.177145+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00412466s) 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: cluster 2026-03-08T23:34:19.177145+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00412466s) 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.181006+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.181006+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.181495+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.181495+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.181840+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.181840+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.182800+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.182800+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.183123+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.183123+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: cluster 2026-03-08T23:34:19.188813+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: cluster 2026-03-08T23:34:19.188813+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.198300+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.198300+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.201092+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.201092+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.206421+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.206421+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.209393+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.209393+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.211614+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:34:19.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:19 vm02 bash[20612]: audit 2026-03-08T23:34:19.211614+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "32c51f16-1b47-11f1-bdef-c3bc34602109", 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-08T23:34:14:089156+0000", 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-08T23:34:20.639 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-08T23:34:14.090156+0000", 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-08T23:34:20.640 INFO:teuthology.orchestra.run.vm02.stdout:mgr is available 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-08T23:34:20.932 INFO:teuthology.orchestra.run.vm02.stdout:Enabling cephadm module... 2026-03-08T23:34:21.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20612]: cluster 2026-03-08T23:34:20.183404+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.01039s) 2026-03-08T23:34:21.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20612]: cluster 2026-03-08T23:34:20.183404+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.01039s) 2026-03-08T23:34:21.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20612]: audit 2026-03-08T23:34:20.585496+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/3641287341' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:34:21.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20612]: audit 2026-03-08T23:34:20.585496+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/3641287341' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:34:21.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20612]: audit 2026-03-08T23:34:20.875686+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1859125897' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-08T23:34:21.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20612]: audit 2026-03-08T23:34:20.875686+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1859125897' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-08T23:34:21.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20612]: audit 2026-03-08T23:34:20.882205+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.102:0/1859125897' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-08T23:34:21.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20612]: audit 2026-03-08T23:34:20.882205+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.102:0/1859125897' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-08T23:34:22.178 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20891]: ignoring --setuser ceph since I am not root 2026-03-08T23:34:22.178 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:21 vm02 bash[20891]: ignoring --setgroup ceph since I am not root 2026-03-08T23:34:22.178 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20891]: debug 2026-03-08T23:34:22.012+0000 7f6faa844140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-08T23:34:22.178 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20891]: debug 2026-03-08T23:34:22.052+0000 7f6faa844140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-08T23:34:22.261 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-08T23:34:22.261 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-08T23:34:22.261 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-08T23:34:22.261 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-08T23:34:22.261 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-08T23:34:22.261 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-08T23:34:22.261 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-08T23:34:22.261 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 4... 2026-03-08T23:34:22.489 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20891]: debug 2026-03-08T23:34:22.176+0000 7f6faa844140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-08T23:34:22.489 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20612]: audit 2026-03-08T23:34:21.200369+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2235398194' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-08T23:34:22.489 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20612]: audit 2026-03-08T23:34:21.200369+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2235398194' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-08T23:34:22.489 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20612]: audit 2026-03-08T23:34:21.882842+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2235398194' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-08T23:34:22.489 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20612]: audit 2026-03-08T23:34:21.882842+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2235398194' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-08T23:34:22.489 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20612]: cluster 2026-03-08T23:34:21.884467+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-08T23:34:22.489 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20612]: cluster 2026-03-08T23:34:21.884467+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-08T23:34:22.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20891]: debug 2026-03-08T23:34:22.484+0000 7f6faa844140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-08T23:34:23.184 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:22 vm02 bash[20891]: debug 2026-03-08T23:34:22.916+0000 7f6faa844140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-08T23:34:23.184 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.000+0000 7f6faa844140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-08T23:34:23.184 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-08T23:34:23.184 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-08T23:34:23.184 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: from numpy import show_config as show_numpy_config 2026-03-08T23:34:23.184 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.116+0000 7f6faa844140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-08T23:34:23.547 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.252+0000 7f6faa844140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-08T23:34:23.547 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.292+0000 7f6faa844140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-08T23:34:23.547 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.328+0000 7f6faa844140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-08T23:34:23.547 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.368+0000 7f6faa844140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-08T23:34:23.547 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.420+0000 7f6faa844140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-08T23:34:23.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20612]: audit 2026-03-08T23:34:22.207526+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/542812958' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-08T23:34:23.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20612]: audit 2026-03-08T23:34:22.207526+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/542812958' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-08T23:34:24.109 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.848+0000 7f6faa844140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-08T23:34:24.109 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.884+0000 7f6faa844140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-08T23:34:24.109 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:23 vm02 bash[20891]: debug 2026-03-08T23:34:23.924+0000 7f6faa844140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-08T23:34:24.110 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.064+0000 7f6faa844140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-08T23:34:24.407 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.108+0000 7f6faa844140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-08T23:34:24.407 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.144+0000 7f6faa844140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-08T23:34:24.407 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.252+0000 7f6faa844140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-08T23:34:24.657 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.404+0000 7f6faa844140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-08T23:34:24.657 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.576+0000 7f6faa844140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-08T23:34:24.657 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.612+0000 7f6faa844140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-08T23:34:25.011 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.656+0000 7f6faa844140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-08T23:34:25.011 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:24 vm02 bash[20891]: debug 2026-03-08T23:34:24.796+0000 7f6faa844140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20891]: debug 2026-03-08T23:34:25.008+0000 7f6faa844140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.014394+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.014394+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.014795+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.014795+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.019732+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.019732+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.019859+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00514881s) 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.019859+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00514881s) 2026-03-08T23:34:25.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.021722+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.021722+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.021809+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.021809+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.022494+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.022494+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.022669+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.022669+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.022833+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.022833+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.027967+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: cluster 2026-03-08T23:34:25.027967+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.037200+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.037200+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.040637+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.040637+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.049287+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.049287+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.052637+0000 mon.a (mon.0) 49 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.052637+0000 mon.a (mon.0) 49 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.054108+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.054108+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.056130+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:34:25.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:25 vm02 bash[20612]: audit 2026-03-08T23:34:25.056130+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:34:26.075 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-08T23:34:26.075 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-08T23:34:26.075 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-08T23:34:26.075 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-08T23:34:26.075 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 4 is available 2026-03-08T23:34:26.075 INFO:teuthology.orchestra.run.vm02.stdout:Setting orchestrator backend to cephadm... 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: cephadm 2026-03-08T23:34:25.034728+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: cephadm 2026-03-08T23:34:25.034728+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:25.611865+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:25.611865+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:25.614474+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:25.614474+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: cluster 2026-03-08T23:34:26.022285+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e6: a(active, since 1.00758s) 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: cluster 2026-03-08T23:34:26.022285+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e6: a(active, since 1.00758s) 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:26.354203+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:26.354203+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:26.358497+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:26.358497+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:26.478662+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:26.622 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20612]: audit 2026-03-08T23:34:26.478662+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:26.647 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-08T23:34:26.647 INFO:teuthology.orchestra.run.vm02.stdout:Generating ssh key... 2026-03-08T23:34:26.887 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: Generating public/private ed25519 key pair. 2026-03-08T23:34:26.887 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: Your identification has been saved in /tmp/tmp87lsy6e7/key 2026-03-08T23:34:26.887 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: Your public key has been saved in /tmp/tmp87lsy6e7/key.pub 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: The key fingerprint is: 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: SHA256:8DrULGhWk6cN7Us/hfEXVs7bq33yXcz+FzcwkImjOiI ceph-32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: The key's randomart image is: 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: +--[ED25519 256]--+ 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | . o .| 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | o o + + | 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | * + o . o o| 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | o & + + .o| 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | + = S . o +..| 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | E + + + o . . =o| 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | . . + . o .B| 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | . . +.=| 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: | . +O| 2026-03-08T23:34:26.888 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:26 vm02 bash[20891]: +----[SHA256]-----+ 2026-03-08T23:34:27.164 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINrX/HUphsBrWNpqf/DYM+YyG3b8wNHp3+N46U+AvkGk ceph-32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:27.164 INFO:teuthology.orchestra.run.vm02.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-08T23:34:27.165 INFO:teuthology.orchestra.run.vm02.stdout:Adding key to root@localhost authorized_keys... 2026-03-08T23:34:27.165 INFO:teuthology.orchestra.run.vm02.stdout:Adding host vm02... 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.024021+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.024021+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.028748+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.028748+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.265165+0000 mgr.a (mgr.14118) 4 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Bus STARTING 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.265165+0000 mgr.a (mgr.14118) 4 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Bus STARTING 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.350763+0000 mgr.a (mgr.14118) 5 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.350763+0000 mgr.a (mgr.14118) 5 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.376552+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Serving on https://192.168.123.102:7150 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.376552+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Serving on https://192.168.123.102:7150 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.377095+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Client ('192.168.123.102', 57968) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.377095+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Client ('192.168.123.102', 57968) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.478023+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Serving on http://192.168.123.102:8765 2026-03-08T23:34:27.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.478023+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Serving on http://192.168.123.102:8765 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.478292+0000 mgr.a (mgr.14118) 9 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Bus STARTED 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.478292+0000 mgr.a (mgr.14118) 9 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Bus STARTED 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.609093+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.609093+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.857623+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.857623+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.857833+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: cephadm 2026-03-08T23:34:26.857833+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.874415+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.874415+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.877531+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:27.990 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:27 vm02 bash[20612]: audit 2026-03-08T23:34:26.877531+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:29.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:28 vm02 bash[20612]: audit 2026-03-08T23:34:27.123391+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:29.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:28 vm02 bash[20612]: audit 2026-03-08T23:34:27.123391+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:29.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:28 vm02 bash[20612]: audit 2026-03-08T23:34:27.410124+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:29.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:28 vm02 bash[20612]: audit 2026-03-08T23:34:27.410124+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:29.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:28 vm02 bash[20612]: cluster 2026-03-08T23:34:27.894773+0000 mon.a (mon.0) 60 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-08T23:34:29.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:28 vm02 bash[20612]: cluster 2026-03-08T23:34:27.894773+0000 mon.a (mon.0) 60 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-08T23:34:29.302 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Added host 'vm02' with addr '192.168.123.102' 2026-03-08T23:34:29.303 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mon service... 2026-03-08T23:34:29.593 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-08T23:34:29.593 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mgr service... 2026-03-08T23:34:29.852 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: cephadm 2026-03-08T23:34:27.952455+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: cephadm 2026-03-08T23:34:27.952455+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: audit 2026-03-08T23:34:29.239240+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: audit 2026-03-08T23:34:29.239240+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: audit 2026-03-08T23:34:29.240101+0000 mon.a (mon.0) 62 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: audit 2026-03-08T23:34:29.240101+0000 mon.a (mon.0) 62 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: audit 2026-03-08T23:34:29.560312+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: audit 2026-03-08T23:34:29.560312+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: audit 2026-03-08T23:34:29.817241+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:30.085 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:29 vm02 bash[20612]: audit 2026-03-08T23:34:29.817241+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:30.368 INFO:teuthology.orchestra.run.vm02.stdout:Enabling the dashboard module... 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: cephadm 2026-03-08T23:34:29.239772+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: cephadm 2026-03-08T23:34:29.239772+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:29.556026+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:29.556026+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: cephadm 2026-03-08T23:34:29.556878+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: cephadm 2026-03-08T23:34:29.556878+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:29.813264+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:29.813264+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: cephadm 2026-03-08T23:34:29.813931+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-08T23:34:31.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: cephadm 2026-03-08T23:34:29.813931+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-08T23:34:31.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:30.073100+0000 mon.a (mon.0) 65 : audit [INF] from='client.? 192.168.123.102:0/2895091647' entity='client.admin' 2026-03-08T23:34:31.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:30.073100+0000 mon.a (mon.0) 65 : audit [INF] from='client.? 192.168.123.102:0/2895091647' entity='client.admin' 2026-03-08T23:34:31.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:30.329602+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.102:0/2893422135' entity='client.admin' 2026-03-08T23:34:31.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:30.329602+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.102:0/2893422135' entity='client.admin' 2026-03-08T23:34:31.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:30.633523+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/1615739362' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-08T23:34:31.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:30.633523+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/1615739362' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-08T23:34:31.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:30.727514+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:31.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:30 vm02 bash[20612]: audit 2026-03-08T23:34:30.727514+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:31.648 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:31 vm02 bash[20891]: ignoring --setuser ceph since I am not root 2026-03-08T23:34:31.648 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:31 vm02 bash[20891]: ignoring --setgroup ceph since I am not root 2026-03-08T23:34:31.648 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:31 vm02 bash[20891]: debug 2026-03-08T23:34:31.488+0000 7f250f564140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-08T23:34:31.648 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:31 vm02 bash[20891]: debug 2026-03-08T23:34:31.520+0000 7f250f564140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-08T23:34:31.757 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-08T23:34:31.757 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-08T23:34:31.757 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-08T23:34:31.757 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-08T23:34:31.757 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-08T23:34:31.757 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-08T23:34:31.757 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-08T23:34:31.757 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 8... 2026-03-08T23:34:31.978 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:31 vm02 bash[20891]: debug 2026-03-08T23:34:31.644+0000 7f250f564140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:31 vm02 bash[20891]: debug 2026-03-08T23:34:31.976+0000 7f250f564140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20612]: audit 2026-03-08T23:34:31.011479+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20612]: audit 2026-03-08T23:34:31.011479+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20612]: audit 2026-03-08T23:34:31.330599+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1615739362' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20612]: audit 2026-03-08T23:34:31.330599+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1615739362' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20612]: cluster 2026-03-08T23:34:31.332841+0000 mon.a (mon.0) 71 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20612]: cluster 2026-03-08T23:34:31.332841+0000 mon.a (mon.0) 71 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20612]: audit 2026-03-08T23:34:31.701419+0000 mon.a (mon.0) 72 : audit [DBG] from='client.? 192.168.123.102:0/714359147' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-08T23:34:32.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20612]: audit 2026-03-08T23:34:31.701419+0000 mon.a (mon.0) 72 : audit [DBG] from='client.? 192.168.123.102:0/714359147' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-08T23:34:32.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: debug 2026-03-08T23:34:32.460+0000 7f250f564140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-08T23:34:32.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: debug 2026-03-08T23:34:32.544+0000 7f250f564140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-08T23:34:32.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-08T23:34:32.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-08T23:34:32.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: from numpy import show_config as show_numpy_config 2026-03-08T23:34:32.797 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: debug 2026-03-08T23:34:32.664+0000 7f250f564140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-08T23:34:33.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: debug 2026-03-08T23:34:32.796+0000 7f250f564140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-08T23:34:33.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: debug 2026-03-08T23:34:32.832+0000 7f250f564140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-08T23:34:33.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: debug 2026-03-08T23:34:32.864+0000 7f250f564140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-08T23:34:33.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: debug 2026-03-08T23:34:32.908+0000 7f250f564140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-08T23:34:33.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:32 vm02 bash[20891]: debug 2026-03-08T23:34:32.956+0000 7f250f564140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-08T23:34:33.642 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:33 vm02 bash[20891]: debug 2026-03-08T23:34:33.384+0000 7f250f564140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-08T23:34:33.642 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:33 vm02 bash[20891]: debug 2026-03-08T23:34:33.420+0000 7f250f564140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-08T23:34:33.642 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:33 vm02 bash[20891]: debug 2026-03-08T23:34:33.456+0000 7f250f564140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-08T23:34:33.642 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:33 vm02 bash[20891]: debug 2026-03-08T23:34:33.600+0000 7f250f564140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-08T23:34:33.942 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:33 vm02 bash[20891]: debug 2026-03-08T23:34:33.640+0000 7f250f564140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-08T23:34:33.942 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:33 vm02 bash[20891]: debug 2026-03-08T23:34:33.676+0000 7f250f564140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-08T23:34:33.942 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:33 vm02 bash[20891]: debug 2026-03-08T23:34:33.788+0000 7f250f564140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-08T23:34:34.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:33 vm02 bash[20891]: debug 2026-03-08T23:34:33.940+0000 7f250f564140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-08T23:34:34.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20891]: debug 2026-03-08T23:34:34.108+0000 7f250f564140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-08T23:34:34.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20891]: debug 2026-03-08T23:34:34.144+0000 7f250f564140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-08T23:34:34.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20891]: debug 2026-03-08T23:34:34.184+0000 7f250f564140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-08T23:34:34.592 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20891]: debug 2026-03-08T23:34:34.324+0000 7f250f564140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-08T23:34:34.592 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20891]: debug 2026-03-08T23:34:34.544+0000 7f250f564140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-08T23:34:35.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.551977+0000 mon.a (mon.0) 73 : cluster [INF] Active manager daemon a restarted 2026-03-08T23:34:35.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.551977+0000 mon.a (mon.0) 73 : cluster [INF] Active manager daemon a restarted 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.552368+0000 mon.a (mon.0) 74 : cluster [INF] Activating manager daemon a 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.552368+0000 mon.a (mon.0) 74 : cluster [INF] Activating manager daemon a 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.557128+0000 mon.a (mon.0) 75 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.557128+0000 mon.a (mon.0) 75 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.557220+0000 mon.a (mon.0) 76 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00494905s) 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.557220+0000 mon.a (mon.0) 76 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00494905s) 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.560072+0000 mon.a (mon.0) 77 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.560072+0000 mon.a (mon.0) 77 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.560373+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.560373+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.563693+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.563693+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.564070+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.564070+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.564436+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.564436+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.570586+0000 mon.a (mon.0) 82 : cluster [INF] Manager daemon a is now available 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: cluster 2026-03-08T23:34:34.570586+0000 mon.a (mon.0) 82 : cluster [INF] Manager daemon a is now available 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.588206+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:35.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:34 vm02 bash[20612]: audit 2026-03-08T23:34:34.588206+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:35.604 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-08T23:34:35.604 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-08T23:34:35.604 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-08T23:34:35.604 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-08T23:34:35.604 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 8 is available 2026-03-08T23:34:35.604 INFO:teuthology.orchestra.run.vm02.stdout:Generating a dashboard self-signed certificate... 2026-03-08T23:34:35.933 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:35 vm02 bash[20612]: audit 2026-03-08T23:34:34.618646+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:34:35.933 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:35 vm02 bash[20612]: audit 2026-03-08T23:34:34.618646+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:34:35.933 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:35 vm02 bash[20612]: audit 2026-03-08T23:34:34.629529+0000 mon.a (mon.0) 85 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:34:35.933 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:35 vm02 bash[20612]: audit 2026-03-08T23:34:34.629529+0000 mon.a (mon.0) 85 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:34:35.933 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:35 vm02 bash[20612]: cluster 2026-03-08T23:34:35.560499+0000 mon.a (mon.0) 86 : cluster [DBG] mgrmap e10: a(active, since 1.00823s) 2026-03-08T23:34:35.933 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:35 vm02 bash[20612]: cluster 2026-03-08T23:34:35.560499+0000 mon.a (mon.0) 86 : cluster [DBG] mgrmap e10: a(active, since 1.00823s) 2026-03-08T23:34:35.959 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-08T23:34:35.959 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial admin user... 2026-03-08T23:34:36.468 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$/I7/yMI8xVwEAbRsBaNT4OuYDEZxsMNKagLxx32HXBD/OAVuS3kdW", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773012876, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-08T23:34:36.468 INFO:teuthology.orchestra.run.vm02.stdout:Fetching dashboard port number... 2026-03-08T23:34:36.727 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 8443 2026-03-08T23:34:36.727 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-08T23:34:36.727 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-08T23:34:36.728 INFO:teuthology.orchestra.run.vm02.stdout:Ceph Dashboard is now available at: 2026-03-08T23:34:36.728 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:36.728 INFO:teuthology.orchestra.run.vm02.stdout: URL: https://vm02.local:8443/ 2026-03-08T23:34:36.728 INFO:teuthology.orchestra.run.vm02.stdout: User: admin 2026-03-08T23:34:36.728 INFO:teuthology.orchestra.run.vm02.stdout: Password: bfwkjp356z 2026-03-08T23:34:36.728 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:36.728 INFO:teuthology.orchestra.run.vm02.stdout:Saving cluster configuration to /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config directory 2026-03-08T23:34:37.024 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.428587+0000 mgr.a (mgr.14150) 1 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Bus STARTING 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.428587+0000 mgr.a (mgr.14150) 1 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Bus STARTING 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.529895+0000 mgr.a (mgr.14150) 2 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Serving on http://192.168.123.102:8765 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.529895+0000 mgr.a (mgr.14150) 2 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Serving on http://192.168.123.102:8765 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.562436+0000 mgr.a (mgr.14150) 3 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.562436+0000 mgr.a (mgr.14150) 3 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.568596+0000 mgr.a (mgr.14150) 4 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.568596+0000 mgr.a (mgr.14150) 4 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.638490+0000 mgr.a (mgr.14150) 5 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Serving on https://192.168.123.102:7150 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.638490+0000 mgr.a (mgr.14150) 5 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Serving on https://192.168.123.102:7150 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.638635+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Bus STARTED 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.638635+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Bus STARTED 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.639019+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Client ('192.168.123.102', 42970) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: cephadm 2026-03-08T23:34:35.639019+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Client ('192.168.123.102', 42970) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.855723+0000 mgr.a (mgr.14150) 8 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.855723+0000 mgr.a (mgr.14150) 8 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.920675+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.920675+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.923742+0000 mon.a (mon.0) 88 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:35.923742+0000 mon.a (mon.0) 88 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:36.201311+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:36.201311+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:36.356947+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:36.356947+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:36.693535+0000 mon.a (mon.0) 90 : audit [DBG] from='client.? 192.168.123.102:0/3855653226' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-08T23:34:37.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:36 vm02 bash[20612]: audit 2026-03-08T23:34:36.693535+0000 mon.a (mon.0) 90 : audit [DBG] from='client.? 192.168.123.102:0/3855653226' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout:Or, if you are only running a single cluster on this host: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: ceph telemetry on 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout:For more information see: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:37.051 INFO:teuthology.orchestra.run.vm02.stdout:Bootstrap complete. 2026-03-08T23:34:37.071 INFO:tasks.cephadm:Fetching config... 2026-03-08T23:34:37.071 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:34:37.071 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-08T23:34:37.073 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-08T23:34:37.073 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:34:37.073 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-08T23:34:37.119 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-08T23:34:37.119 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:34:37.119 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/keyring of=/dev/stdout 2026-03-08T23:34:37.167 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-08T23:34:37.167 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:34:37.167 DEBUG:teuthology.orchestra.run.vm02:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-08T23:34:37.211 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-08T23:34:37.211 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINrX/HUphsBrWNpqf/DYM+YyG3b8wNHp3+N46U+AvkGk ceph-32c51f16-1b47-11f1-bdef-c3bc34602109' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-08T23:34:37.264 INFO:teuthology.orchestra.run.vm02.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINrX/HUphsBrWNpqf/DYM+YyG3b8wNHp3+N46U+AvkGk ceph-32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:37.270 DEBUG:teuthology.orchestra.run.vm05:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINrX/HUphsBrWNpqf/DYM+YyG3b8wNHp3+N46U+AvkGk ceph-32c51f16-1b47-11f1-bdef-c3bc34602109' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-08T23:34:37.281 INFO:teuthology.orchestra.run.vm05.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINrX/HUphsBrWNpqf/DYM+YyG3b8wNHp3+N46U+AvkGk ceph-32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:34:37.285 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-08T23:34:38.016 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:38 vm02 bash[20612]: audit 2026-03-08T23:34:37.017094+0000 mon.a (mon.0) 91 : audit [INF] from='client.? 192.168.123.102:0/3679707835' entity='client.admin' 2026-03-08T23:34:38.017 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:38 vm02 bash[20612]: audit 2026-03-08T23:34:37.017094+0000 mon.a (mon.0) 91 : audit [INF] from='client.? 192.168.123.102:0/3679707835' entity='client.admin' 2026-03-08T23:34:38.017 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:38 vm02 bash[20612]: cluster 2026-03-08T23:34:37.362094+0000 mon.a (mon.0) 92 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-08T23:34:38.017 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:38 vm02 bash[20612]: cluster 2026-03-08T23:34:37.362094+0000 mon.a (mon.0) 92 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-08T23:34:40.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:39 vm02 bash[20612]: audit 2026-03-08T23:34:38.906681+0000 mon.a (mon.0) 93 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:40.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:39 vm02 bash[20612]: audit 2026-03-08T23:34:38.906681+0000 mon.a (mon.0) 93 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:40.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:39 vm02 bash[20612]: audit 2026-03-08T23:34:39.458039+0000 mon.a (mon.0) 94 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:40.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:39 vm02 bash[20612]: audit 2026-03-08T23:34:39.458039+0000 mon.a (mon.0) 94 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:41.781 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:34:42.082 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:41 vm02 bash[20612]: cluster 2026-03-08T23:34:40.913679+0000 mon.a (mon.0) 95 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-08T23:34:42.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:41 vm02 bash[20612]: cluster 2026-03-08T23:34:40.913679+0000 mon.a (mon.0) 95 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-08T23:34:42.123 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-08T23:34:42.123 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-08T23:34:43.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:43 vm02 bash[20612]: audit 2026-03-08T23:34:42.069599+0000 mon.a (mon.0) 96 : audit [INF] from='client.? 192.168.123.102:0/18452878' entity='client.admin' 2026-03-08T23:34:43.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:43 vm02 bash[20612]: audit 2026-03-08T23:34:42.069599+0000 mon.a (mon.0) 96 : audit [INF] from='client.? 192.168.123.102:0/18452878' entity='client.admin' 2026-03-08T23:34:45.790 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:34:46.102 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm05 2026-03-08T23:34:46.103 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:34:46.103 DEBUG:teuthology.orchestra.run.vm05:> dd of=/etc/ceph/ceph.conf 2026-03-08T23:34:46.105 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:34:46.106 DEBUG:teuthology.orchestra.run.vm05:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:34:46.148 INFO:tasks.cephadm:Adding host vm05 to orchestrator... 2026-03-08T23:34:46.148 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch host add vm05 2026-03-08T23:34:46.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.291668+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.291668+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.293959+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.293959+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.294502+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.294502+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.296848+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.296848+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.302185+0000 mon.a (mon.0) 101 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.302185+0000 mon.a (mon.0) 101 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.304530+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:45.304530+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.027051+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.027051+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.027611+0000 mon.a (mon.0) 104 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.027611+0000 mon.a (mon.0) 104 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.028620+0000 mon.a (mon.0) 105 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.028620+0000 mon.a (mon.0) 105 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.029059+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.029059+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.165836+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.165836+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.169182+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.169182+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.172856+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:46.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:46 vm02 bash[20612]: audit 2026-03-08T23:34:46.172856+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:47.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: audit 2026-03-08T23:34:46.024136+0000 mgr.a (mgr.14150) 10 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:47.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: audit 2026-03-08T23:34:46.024136+0000 mgr.a (mgr.14150) 10 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:47.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: cephadm 2026-03-08T23:34:46.029702+0000 mgr.a (mgr.14150) 11 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-08T23:34:47.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: cephadm 2026-03-08T23:34:46.029702+0000 mgr.a (mgr.14150) 11 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-08T23:34:47.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: cephadm 2026-03-08T23:34:46.065452+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:34:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: cephadm 2026-03-08T23:34:46.065452+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:34:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: cephadm 2026-03-08T23:34:46.105796+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:34:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: cephadm 2026-03-08T23:34:46.105796+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:34:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: cephadm 2026-03-08T23:34:46.134027+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.client.admin.keyring 2026-03-08T23:34:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:47 vm02 bash[20612]: cephadm 2026-03-08T23:34:46.134027+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.client.admin.keyring 2026-03-08T23:34:49.797 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:34:50.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:50 vm02 bash[20612]: audit 2026-03-08T23:34:50.063273+0000 mgr.a (mgr.14150) 15 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:50.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:50 vm02 bash[20612]: audit 2026-03-08T23:34:50.063273+0000 mgr.a (mgr.14150) 15 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:34:51.790 INFO:teuthology.orchestra.run.vm02.stdout:Added host 'vm05' with addr '192.168.123.105' 2026-03-08T23:34:51.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:51 vm02 bash[20612]: cephadm 2026-03-08T23:34:50.591524+0000 mgr.a (mgr.14150) 16 : cephadm [INF] Deploying cephadm binary to vm05 2026-03-08T23:34:51.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:51 vm02 bash[20612]: cephadm 2026-03-08T23:34:50.591524+0000 mgr.a (mgr.14150) 16 : cephadm [INF] Deploying cephadm binary to vm05 2026-03-08T23:34:51.847 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch host ls --format=json 2026-03-08T23:34:53.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:52 vm02 bash[20612]: audit 2026-03-08T23:34:51.791523+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:53.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:52 vm02 bash[20612]: audit 2026-03-08T23:34:51.791523+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:53.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:52 vm02 bash[20612]: cephadm 2026-03-08T23:34:51.791857+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Added host vm05 2026-03-08T23:34:53.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:52 vm02 bash[20612]: cephadm 2026-03-08T23:34:51.791857+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Added host vm05 2026-03-08T23:34:53.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:52 vm02 bash[20612]: audit 2026-03-08T23:34:51.792341+0000 mon.a (mon.0) 111 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:53.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:52 vm02 bash[20612]: audit 2026-03-08T23:34:51.792341+0000 mon.a (mon.0) 111 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:34:53.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:52 vm02 bash[20612]: audit 2026-03-08T23:34:52.093729+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:53.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:52 vm02 bash[20612]: audit 2026-03-08T23:34:52.093729+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:54.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:54 vm02 bash[20612]: audit 2026-03-08T23:34:53.376093+0000 mon.a (mon.0) 113 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:54.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:54 vm02 bash[20612]: audit 2026-03-08T23:34:53.376093+0000 mon.a (mon.0) 113 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:54.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:54 vm02 bash[20612]: audit 2026-03-08T23:34:53.920770+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:54.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:54 vm02 bash[20612]: audit 2026-03-08T23:34:53.920770+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:55 vm02 bash[20612]: cluster 2026-03-08T23:34:54.564624+0000 mgr.a (mgr.14150) 18 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:34:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:55 vm02 bash[20612]: cluster 2026-03-08T23:34:54.564624+0000 mgr.a (mgr.14150) 18 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:34:56.460 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:34:56.721 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:34:56.721 INFO:teuthology.orchestra.run.vm02.stdout:[{"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.105", "hostname": "vm05", "labels": [], "status": ""}] 2026-03-08T23:34:56.769 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-08T23:34:56.769 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph osd crush tunables default 2026-03-08T23:34:57.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cluster 2026-03-08T23:34:56.564791+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cluster 2026-03-08T23:34:56.564791+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.644185+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.644185+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.646644+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.646644+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.649203+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.649203+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.651621+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.651621+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.652213+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.652213+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.652816+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.652816+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.653186+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.653186+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cephadm 2026-03-08T23:34:56.653726+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cephadm 2026-03-08T23:34:56.653726+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cephadm 2026-03-08T23:34:56.685014+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cephadm 2026-03-08T23:34:56.685014+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cephadm 2026-03-08T23:34:56.712130+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cephadm 2026-03-08T23:34:56.712130+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.723095+0000 mgr.a (mgr.14150) 23 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.723095+0000 mgr.a (mgr.14150) 23 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cephadm 2026-03-08T23:34:56.762975+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.client.admin.keyring 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: cephadm 2026-03-08T23:34:56.762975+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.client.admin.keyring 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.802849+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.802849+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.805009+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.805009+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.807099+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:34:57.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:57 vm02 bash[20612]: audit 2026-03-08T23:34:56.807099+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:00.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:59 vm02 bash[20612]: cluster 2026-03-08T23:34:58.564958+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:00.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:34:59 vm02 bash[20612]: cluster 2026-03-08T23:34:58.564958+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:00.468 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:35:01.652 INFO:teuthology.orchestra.run.vm02.stderr:adjusted tunables profile to default 2026-03-08T23:35:01.704 INFO:tasks.cephadm:Adding mon.a on vm02 2026-03-08T23:35:01.704 INFO:tasks.cephadm:Adding mon.b on vm05 2026-03-08T23:35:01.704 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch apply mon '2;vm02:192.168.123.102=a;vm05:192.168.123.105=b' 2026-03-08T23:35:01.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:01 vm02 bash[20612]: cluster 2026-03-08T23:35:00.565067+0000 mgr.a (mgr.14150) 26 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:01.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:01 vm02 bash[20612]: cluster 2026-03-08T23:35:00.565067+0000 mgr.a (mgr.14150) 26 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:01.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:01 vm02 bash[20612]: audit 2026-03-08T23:35:00.710298+0000 mon.a (mon.0) 125 : audit [INF] from='client.? 192.168.123.102:0/1767454902' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-08T23:35:01.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:01 vm02 bash[20612]: audit 2026-03-08T23:35:00.710298+0000 mon.a (mon.0) 125 : audit [INF] from='client.? 192.168.123.102:0/1767454902' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-08T23:35:02.812 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:03.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:02 vm02 bash[20612]: audit 2026-03-08T23:35:01.653840+0000 mon.a (mon.0) 126 : audit [INF] from='client.? 192.168.123.102:0/1767454902' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-08T23:35:03.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:02 vm02 bash[20612]: audit 2026-03-08T23:35:01.653840+0000 mon.a (mon.0) 126 : audit [INF] from='client.? 192.168.123.102:0/1767454902' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-08T23:35:03.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:02 vm02 bash[20612]: cluster 2026-03-08T23:35:01.655578+0000 mon.a (mon.0) 127 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-08T23:35:03.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:02 vm02 bash[20612]: cluster 2026-03-08T23:35:01.655578+0000 mon.a (mon.0) 127 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-08T23:35:03.060 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled mon update... 2026-03-08T23:35:03.138 DEBUG:teuthology.orchestra.run.vm05:mon.b> sudo journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.b.service 2026-03-08T23:35:03.139 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-08T23:35:03.139 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph mon dump -f json 2026-03-08T23:35:04.287 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.b/config 2026-03-08T23:35:04.540 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 systemd[1]: Started Ceph mon.b for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: cluster 2026-03-08T23:35:02.565293+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: cluster 2026-03-08T23:35:02.565293+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.057652+0000 mgr.a (mgr.14150) 28 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm02:192.168.123.102=a;vm05:192.168.123.105=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.057652+0000 mgr.a (mgr.14150) 28 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm02:192.168.123.102=a;vm05:192.168.123.105=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: cephadm 2026-03-08T23:35:03.058692+0000 mgr.a (mgr.14150) 29 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm05:192.168.123.105=b;count:2 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: cephadm 2026-03-08T23:35:03.058692+0000 mgr.a (mgr.14150) 29 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm05:192.168.123.105=b;count:2 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.061386+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.061386+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.062051+0000 mon.a (mon.0) 129 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.062051+0000 mon.a (mon.0) 129 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.063178+0000 mon.a (mon.0) 130 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.063178+0000 mon.a (mon.0) 130 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.063632+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.063632+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.066791+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.066791+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.067696+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:04.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.067696+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:04.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.068081+0000 mon.a (mon.0) 134 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:04.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: audit 2026-03-08T23:35:03.068081+0000 mon.a (mon.0) 134 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:04.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: cephadm 2026-03-08T23:35:03.068599+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-08T23:35:04.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:04 vm02 bash[20612]: cephadm 2026-03-08T23:35:03.068599+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 0 pidfile_write: ignore empty --pid-file 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 0 load: jerasure load: lrc 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Git sha 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: DB SUMMARY 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: DB Session ID: PPAGD12ZYWB9ZD1OFUG6 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: CURRENT file: CURRENT 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 0, files: 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000004.log size: 511 ; 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.error_if_exists: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.create_if_missing: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.env: 0x56183432adc0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.info_log: 0x56186b529880 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.statistics: (nil) 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.use_fsync: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.db_log_dir: 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.wal_dir: 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.write_buffer_manager: 0x56186b52d900 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-08T23:35:04.794 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.unordered_write: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.row_cache: None 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.wal_filter: None 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.two_write_queues: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.wal_compression: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.atomic_flush: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_open_files: -1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Compression algorithms supported: 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: kZSTD supported: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: kXpressCompression supported: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: kZlibCompression supported: 1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-08T23:35:04.795 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.merge_operator: 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_filter: None 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x56186b5294e0) 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cache_index_and_filter_blocks: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: pin_top_level_index_and_filter: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: index_type: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: data_block_index_type: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: index_shortening: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: data_block_hash_table_util_ratio: 0.750000 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: checksum: 4 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: no_block_cache: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: block_cache: 0x56186b54e9b0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: block_cache_name: BinnedLRUCache 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: block_cache_options: 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: capacity : 536870912 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: num_shard_bits : 4 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: strict_capacity_limit : 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: high_pri_pool_ratio: 0.000 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: block_cache_compressed: (nil) 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: persistent_cache: (nil) 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: block_size: 4096 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: block_size_deviation: 10 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: block_restart_interval: 16 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: index_block_restart_interval: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: metadata_block_size: 4096 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: partition_filters: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: use_delta_encoding: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: filter_policy: bloomfilter 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: whole_key_filtering: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: verify_compression: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: read_amp_bytes_per_bit: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: format_version: 5 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: enable_index_compression: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: block_align: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: max_auto_readahead_size: 262144 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: prepopulate_block_cache: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: initial_auto_readahead_size: 8192 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: num_file_reads_for_auto_readahead: 2 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression: NoCompression 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.num_levels: 7 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-08T23:35:04.796 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.bloom_locality: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.ttl: 2592000 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.enable_blob_files: false 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.min_blob_size: 0 2026-03-08T23:35:04.797 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 7ce5a3f0-68c5-45fd-bad2-ce95fe1c5ad6 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773012904749700, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.745+0000 7fe279506d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.753+0000 7fe279506d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773012904756841, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773012904, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "7ce5a3f0-68c5-45fd-bad2-ce95fe1c5ad6", "db_session_id": "PPAGD12ZYWB9ZD1OFUG6", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.753+0000 7fe279506d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773012904756897, "job": 1, "event": "recovery_finished"} 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.753+0000 7fe279506d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.757+0000 7fe279506d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.757+0000 7fe279506d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x56186b550e00 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.757+0000 7fe279506d80 4 rocksdb: DB pointer 0x56186b660000 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.757+0000 7fe279506d80 0 mon.b does not exist in monmap, will attempt to join an existing cluster 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.757+0000 7fe279506d80 0 using public_addr v2:192.168.123.105:0/0 -> [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.757+0000 7fe279506d80 0 starting mon.b rank -1 at public addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] at bind addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.761+0000 7fe279506d80 1 mon.b@-1(???) e0 preinit fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.761+0000 7fe26f2d0640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.761+0000 7fe26f2d0640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: ** DB Stats ** 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: ** Compaction Stats [default] ** 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.007 0 0 0.0 0.0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.007 0 0 0.0 0.0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.007 0 0 0.0 0.0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: ** Compaction Stats [default] ** 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.2 0.01 0.00 1 0.007 0 0 0.0 0.0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: AddFile(Total Files): cumulative 0, interval 0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: AddFile(Keys): cumulative 0, interval 0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Cumulative compaction: 0.00 GB write, 0.11 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Interval compaction: 0.00 GB write, 0.11 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Block cache BinnedLRUCache@0x56186b54e9b0#7 capacity: 512.00 MB usage: 0.86 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 6e-06 secs_since: 0 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: Block cache entry stats(count,size,portion): DataBlock(1,0.64 KB,0.00012219%) FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: ** File Read Latency Histogram By Level [default] ** 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 0 mon.b@-1(synchronizing).mds e1 new map 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 0 mon.b@-1(synchronizing).mds e1 print_map 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: e1 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: btime 2026-03-08T23:34:14:089156+0000 2026-03-08T23:35:04.798 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: legacy client fscid: -1 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: No filesystems configured 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 1 mon.b@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 1 mon.b@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:14.089854+0000 mon.a (mon.0) 0 : cluster [INF] mkfs 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 1 mon.b@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:14.089854+0000 mon.a (mon.0) 0 : cluster [INF] mkfs 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:14.075303+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:14.075303+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191286+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191286+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191342+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191342+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191349+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191349+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191354+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-08T23:34:12.859532+0000 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191354+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-08T23:34:12.859532+0000 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191363+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191363+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191368+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191368+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191374+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191374+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191379+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191379+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191699+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191699+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191718+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.191718+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.192321+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:15.192321+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:15.436282+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1454493050' entity='client.admin' 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:15.436282+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1454493050' entity='client.admin' 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:16.038600+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4188746370' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:16.038600+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4188746370' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:18.278146+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/2262746306' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:18.278146+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/2262746306' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:19.173138+0000 mon.a (mon.0) 15 : clusterdebug 2026-03-08T23:35:04.785+0000 7fe2722d6640 1 mon.b@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 1 mon.b@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 1 mon.b@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: [INF] Activating manager daemon a 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:19.173138+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:19.177145+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00412466s) 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:19.177145+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00412466s) 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.181006+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.181006+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.181495+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.181495+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.181840+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.181840+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.182800+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.182800+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.183123+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.183123+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:19.188813+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:19.188813+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.198300+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.198300+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.201092+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:35:04.799 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.201092+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.206421+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.206421+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.209393+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.209393+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.211614+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:19.211614+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14102 192.168.123.102:0/960153252' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:20.183404+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.01039s) 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:20.183404+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.01039s) 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:20.585496+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/3641287341' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:20.585496+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/3641287341' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:20.875686+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1859125897' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:20.875686+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1859125897' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:20.882205+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.102:0/1859125897' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:20.882205+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.102:0/1859125897' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:21.200369+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2235398194' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:21.200369+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2235398194' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:21.882842+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2235398194' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:21.882842+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2235398194' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:21.884467+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:21.884467+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:22.207526+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/542812958' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:22.207526+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/542812958' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.014394+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.014394+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.014795+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.014795+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.019732+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.019732+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.019859+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00514881s) 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.019859+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00514881s) 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.021722+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.021722+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.021809+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.021809+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.022494+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.022494+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.022669+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.022669+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.022833+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.022833+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.027967+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:25.027967+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.037200+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.037200+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.040637+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.040637+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.049287+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.049287+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.052637+0000 mon.a (mon.0) 49 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.052637+0000 mon.a (mon.0) 49 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.054108+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.054108+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.056130+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.056130+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:25.034728+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:25.034728+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.611865+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.611865+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.614474+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:25.614474+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:26.022285+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e6: a(active, since 1.00758s) 2026-03-08T23:35:04.800 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:26.022285+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e6: a(active, since 1.00758s) 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.354203+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.354203+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.358497+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.358497+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.478662+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.478662+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.024021+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.024021+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.028748+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.028748+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.265165+0000 mgr.a (mgr.14118) 4 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Bus STARTING 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.265165+0000 mgr.a (mgr.14118) 4 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Bus STARTING 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.350763+0000 mgr.a (mgr.14118) 5 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.350763+0000 mgr.a (mgr.14118) 5 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.376552+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Serving on https://192.168.123.102:7150 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.376552+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Serving on https://192.168.123.102:7150 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.377095+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Client ('192.168.123.102', 57968) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.377095+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Client ('192.168.123.102', 57968) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.478023+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Serving on http://192.168.123.102:8765 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.478023+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Serving on http://192.168.123.102:8765 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.478292+0000 mgr.a (mgr.14118) 9 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Bus STARTED 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.478292+0000 mgr.a (mgr.14118) 9 : cephadm [INF] [08/Mar/2026:23:34:26] ENGINE Bus STARTED 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.609093+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.609093+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.857623+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.857623+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.857833+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:26.857833+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.874415+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.874415+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.877531+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:26.877531+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:27.123391+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:27.123391+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:27.410124+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:27.410124+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:27.894773+0000 mon.a (mon.0) 60 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:27.894773+0000 mon.a (mon.0) 60 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:27.952455+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:27.952455+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.239240+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.239240+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.240101+0000 mon.a (mon.0) 62 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.240101+0000 mon.a (mon.0) 62 : audit [DBG] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.560312+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.560312+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.817241+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.817241+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:29.239772+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:29.239772+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.556026+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.556026+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.801 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:29.556878+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:29.556878+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.813264+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:29.813264+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:29.813931+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:29.813931+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:30.073100+0000 mon.a (mon.0) 65 : audit [INF] from='client.? 192.168.123.102:0/2895091647' entity='client.admin' 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:30.073100+0000 mon.a (mon.0) 65 : audit [INF] from='client.? 192.168.123.102:0/2895091647' entity='client.admin' 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:30.329602+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.102:0/2893422135' entity='client.admin' 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:30.329602+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.102:0/2893422135' entity='client.admin' 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:30.633523+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/1615739362' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:30.633523+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/1615739362' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:30.727514+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:30.727514+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:31.011479+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:31.011479+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/3999504132' entity='mgr.a' 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:31.330599+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1615739362' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:31.330599+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1615739362' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:31.332841+0000 mon.a (mon.0) 71 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:31.332841+0000 mon.a (mon.0) 71 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:31.701419+0000 mon.a (mon.0) 72 : audit [DBG] from='client.? 192.168.123.102:0/714359147' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:31.701419+0000 mon.a (mon.0) 72 : audit [DBG] from='client.? 192.168.123.102:0/714359147' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 0 mon.b@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: clusterdebug 2026-03-08T23:35:04.785+0000 7fe2722d6640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: 2026debug 2026-03-08T23:35:04.785+0000 7fe2722d6640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: -03-08T23:34:34.551977+0000 mon.a (mon.0) 73 : cluster [INF] Active manager daemon a restarted 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.551977+0000 mon.a (mon.0) 73 : cluster [INF] Active manager daemon a restarted 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.552368+0000 mon.a (mon.0) 74 : cluster [INF] Activating manager daemon a 2026-03-08T23:35:04.802 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.552368+0000 mon.a (mon.0) 74 : cluster [INF] Activating manager daemon a 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.557128+0000 mon.a (mon.0) 75 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.557128+0000 mon.a (mon.0) 75 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.557220+0000 mon.a (mon.0) 76 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00494905s) 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.557220+0000 mon.a (mon.0) 76 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00494905s) 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.560072+0000 mon.a (mon.0) 77 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.560072+0000 mon.a (mon.0) 77 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.560373+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.560373+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.563693+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.563693+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.564070+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.564070+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.564436+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.564436+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.570586+0000 mon.a (mon.0) 82 : cluster [INF] Manager daemon a is now available 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:34.570586+0000 mon.a (mon.0) 82 : cluster [INF] Manager daemon a is now available 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.588206+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.588206+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.618646+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.618646+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.629529+0000 mon.a (mon.0) 85 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:34.629529+0000 mon.a (mon.0) 85 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:35.560499+0000 mon.a (mon.0) 86 : cluster [DBG] mgrmap e10: a(active, since 1.00823s) 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:35.560499+0000 mon.a (mon.0) 86 : cluster [DBG] mgrmap e10: a(active, since 1.00823s) 2026-03-08T23:35:05.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.428587+0000 mgr.a (mgr.14150) 1 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Bus STARTING 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.428587+0000 mgr.a (mgr.14150) 1 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Bus STARTING 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.529895+0000 mgr.a (mgr.14150) 2 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Serving on http://192.168.123.102:8765 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.529895+0000 mgr.a (mgr.14150) 2 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Serving on http://192.168.123.102:8765 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.562436+0000 mgr.a (mgr.14150) 3 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.562436+0000 mgr.a (mgr.14150) 3 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.568596+0000 mgr.a (mgr.14150) 4 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.568596+0000 mgr.a (mgr.14150) 4 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.638490+0000 mgr.a (mgr.14150) 5 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Serving on https://192.168.123.102:7150 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.638490+0000 mgr.a (mgr.14150) 5 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Serving on https://192.168.123.102:7150 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.638635+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Bus STARTED 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.638635+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Bus STARTED 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.639019+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Client ('192.168.123.102', 42970) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:35.639019+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [08/Mar/2026:23:34:35] ENGINE Client ('192.168.123.102', 42970) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.855723+0000 mgr.a (mgr.14150) 8 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.855723+0000 mgr.a (mgr.14150) 8 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.920675+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.920675+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.923742+0000 mon.a (mon.0) 88 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:35.923742+0000 mon.a (mon.0) 88 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:36.201311+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:36.201311+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:36.356947+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:36.356947+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:36.693535+0000 mon.a (mon.0) 90 : audit [DBG] from='client.? 192.168.123.102:0/3855653226' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:36.693535+0000 mon.a (mon.0) 90 : audit [DBG] from='client.? 192.168.123.102:0/3855653226' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:37.017094+0000 mon.a (mon.0) 91 : audit [INF] from='client.? 192.168.123.102:0/3679707835' entity='client.admin' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:37.017094+0000 mon.a (mon.0) 91 : audit [INF] from='client.? 192.168.123.102:0/3679707835' entity='client.admin' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:37.362094+0000 mon.a (mon.0) 92 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:37.362094+0000 mon.a (mon.0) 92 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:38.906681+0000 mon.a (mon.0) 93 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:38.906681+0000 mon.a (mon.0) 93 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:39.458039+0000 mon.a (mon.0) 94 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:39.458039+0000 mon.a (mon.0) 94 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:40.913679+0000 mon.a (mon.0) 95 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:40.913679+0000 mon.a (mon.0) 95 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:42.069599+0000 mon.a (mon.0) 96 : audit [INF] from='client.? 192.168.123.102:0/18452878' entity='client.admin' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:42.069599+0000 mon.a (mon.0) 96 : audit [INF] from='client.? 192.168.123.102:0/18452878' entity='client.admin' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.291668+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.291668+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.293959+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.293959+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.294502+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.294502+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.296848+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.296848+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.302185+0000 mon.a (mon.0) 101 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.302185+0000 mon.a (mon.0) 101 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.304530+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:45.304530+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.027051+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.027051+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.027611+0000 mon.a (mon.0) 104 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.027611+0000 mon.a (mon.0) 104 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.028620+0000 mon.a (mon.0) 105 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.028620+0000 mon.a (mon.0) 105 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.029059+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.029059+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:05.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.165836+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.165836+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.169182+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.169182+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.172856+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.172856+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.024136+0000 mgr.a (mgr.14150) 10 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:46.024136+0000 mgr.a (mgr.14150) 10 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:46.029702+0000 mgr.a (mgr.14150) 11 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:46.029702+0000 mgr.a (mgr.14150) 11 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:46.065452+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:46.065452+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:46.105796+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:46.105796+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:46.134027+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.client.admin.keyring 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:46.134027+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.client.admin.keyring 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:50.063273+0000 mgr.a (mgr.14150) 15 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:50.063273+0000 mgr.a (mgr.14150) 15 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:50.591524+0000 mgr.a (mgr.14150) 16 : cephadm [INF] Deploying cephadm binary to vm05 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:50.591524+0000 mgr.a (mgr.14150) 16 : cephadm [INF] Deploying cephadm binary to vm05 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:51.791523+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:51.791523+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:51.791857+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Added host vm05 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:51.791857+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Added host vm05 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:51.792341+0000 mon.a (mon.0) 111 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:51.792341+0000 mon.a (mon.0) 111 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:52.093729+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:52.093729+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:53.376093+0000 mon.a (mon.0) 113 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:53.376093+0000 mon.a (mon.0) 113 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:53.920770+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:53.920770+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:54.564624+0000 mgr.a (mgr.14150) 18 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:54.564624+0000 mgr.a (mgr.14150) 18 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:56.564791+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:56.564791+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.644185+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.644185+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.646644+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.646644+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.649203+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.649203+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.651621+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.651621+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.652213+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.652213+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.652816+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.652816+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.653186+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.653186+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:05.252 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:56.653726+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:56.653726+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:56.685014+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:56.685014+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:56.712130+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:56.712130+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.723095+0000 mgr.a (mgr.14150) 23 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.723095+0000 mgr.a (mgr.14150) 23 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:56.762975+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.client.admin.keyring 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:34:56.762975+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.client.admin.keyring 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.802849+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.802849+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.805009+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.805009+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.807099+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:34:56.807099+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:58.564958+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:34:58.564958+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:35:00.565067+0000 mgr.a (mgr.14150) 26 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:35:00.565067+0000 mgr.a (mgr.14150) 26 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:00.710298+0000 mon.a (mon.0) 125 : audit [INF] from='client.? 192.168.123.102:0/1767454902' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:00.710298+0000 mon.a (mon.0) 125 : audit [INF] from='client.? 192.168.123.102:0/1767454902' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:01.653840+0000 mon.a (mon.0) 126 : audit [INF] from='client.? 192.168.123.102:0/1767454902' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:01.653840+0000 mon.a (mon.0) 126 : audit [INF] from='client.? 192.168.123.102:0/1767454902' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:35:01.655578+0000 mon.a (mon.0) 127 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:35:01.655578+0000 mon.a (mon.0) 127 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:35:02.565293+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cluster 2026-03-08T23:35:02.565293+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.057652+0000 mgr.a (mgr.14150) 28 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm02:192.168.123.102=a;vm05:192.168.123.105=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.057652+0000 mgr.a (mgr.14150) 28 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm02:192.168.123.102=a;vm05:192.168.123.105=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:35:03.058692+0000 mgr.a (mgr.14150) 29 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm05:192.168.123.105=b;count:2 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:35:03.058692+0000 mgr.a (mgr.14150) 29 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm05:192.168.123.105=b;count:2 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.061386+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.061386+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.062051+0000 mon.a (mon.0) 129 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.062051+0000 mon.a (mon.0) 129 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.063178+0000 mon.a (mon.0) 130 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.063178+0000 mon.a (mon.0) 130 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.063632+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.063632+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.066791+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.066791+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.067696+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.067696+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.068081+0000 mon.a (mon.0) 134 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: audit 2026-03-08T23:35:03.068081+0000 mon.a (mon.0) 134 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:35:03.068599+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-08T23:35:05.253 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:04 vm05 bash[22859]: cephadm 2026-03-08T23:35:03.068599+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying daemon mon.b on vm05 2026-03-08T23:35:09.821 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:35:09.822 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":2,"fsid":"32c51f16-1b47-11f1-bdef-c3bc34602109","modified":"2026-03-08T23:35:04.807804Z","created":"2026-03-08T23:34:12.859532Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:3300","nonce":0},{"type":"v1","addr":"192.168.123.105:6789","nonce":0}]},"addr":"192.168.123.105:6789/0","public_addr":"192.168.123.105:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-08T23:35:09.822 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 2 2026-03-08T23:35:09.898 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-08T23:35:09.898 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph config generate-minimal-conf 2026-03-08T23:35:10.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:04.565465+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:04.565465+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:04.811652+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:04.811652+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:04.811859+0000 mon.a (mon.0) 143 : cluster [INF] mon.a calling monitor election 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:04.811859+0000 mon.a (mon.0) 143 : cluster [INF] mon.a calling monitor election 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:04.813049+0000 mon.a (mon.0) 144 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:04.813049+0000 mon.a (mon.0) 144 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:04.829590+0000 mon.a (mon.0) 145 : audit [DBG] from='client.? 192.168.123.105:0/475018937' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:04.829590+0000 mon.a (mon.0) 145 : audit [DBG] from='client.? 192.168.123.105:0/475018937' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:05.802063+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:05.802063+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:06.565620+0000 mgr.a (mgr.14150) 32 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:06.565620+0000 mgr.a (mgr.14150) 32 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:06.801242+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:06.801242+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:06.802509+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:06.802509+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:07.802108+0000 mon.a (mon.0) 148 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:07.802108+0000 mon.a (mon.0) 148 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:08.565786+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:08.565786+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:08.802288+0000 mon.a (mon.0) 149 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:08.802288+0000 mon.a (mon.0) 149 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.802330+0000 mon.a (mon.0) 150 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.802330+0000 mon.a (mon.0) 150 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.815964+0000 mon.a (mon.0) 151 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.815964+0000 mon.a (mon.0) 151 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819666+0000 mon.a (mon.0) 152 : cluster [DBG] monmap epoch 2 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819666+0000 mon.a (mon.0) 152 : cluster [DBG] monmap epoch 2 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819717+0000 mon.a (mon.0) 153 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819717+0000 mon.a (mon.0) 153 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819758+0000 mon.a (mon.0) 154 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819758+0000 mon.a (mon.0) 154 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819798+0000 mon.a (mon.0) 155 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819798+0000 mon.a (mon.0) 155 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819838+0000 mon.a (mon.0) 156 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819838+0000 mon.a (mon.0) 156 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819879+0000 mon.a (mon.0) 157 : cluster [DBG] election_strategy: 1 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819879+0000 mon.a (mon.0) 157 : cluster [DBG] election_strategy: 1 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819919+0000 mon.a (mon.0) 158 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819919+0000 mon.a (mon.0) 158 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819964+0000 mon.a (mon.0) 159 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.819964+0000 mon.a (mon.0) 159 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.820361+0000 mon.a (mon.0) 160 : cluster [DBG] fsmap 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.820361+0000 mon.a (mon.0) 160 : cluster [DBG] fsmap 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.820456+0000 mon.a (mon.0) 161 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.820456+0000 mon.a (mon.0) 161 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.820685+0000 mon.a (mon.0) 162 : cluster [DBG] mgrmap e12: a(active, since 35s) 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.820685+0000 mon.a (mon.0) 162 : cluster [DBG] mgrmap e12: a(active, since 35s) 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.820961+0000 mon.a (mon.0) 163 : cluster [INF] overall HEALTH_OK 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: cluster 2026-03-08T23:35:09.820961+0000 mon.a (mon.0) 163 : cluster [INF] overall HEALTH_OK 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.826605+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.826605+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.831281+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.831281+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:10.218 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.832056+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:10.219 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.832056+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:10.219 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.832922+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:10.219 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:09 vm02 bash[20612]: audit 2026-03-08T23:35:09.832922+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:10.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:04.565465+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:04.565465+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:04.811652+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:10.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:04.811652+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:35:10.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:04.811859+0000 mon.a (mon.0) 143 : cluster [INF] mon.a calling monitor election 2026-03-08T23:35:10.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:04.811859+0000 mon.a (mon.0) 143 : cluster [INF] mon.a calling monitor election 2026-03-08T23:35:10.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:04.813049+0000 mon.a (mon.0) 144 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:04.813049+0000 mon.a (mon.0) 144 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:04.829590+0000 mon.a (mon.0) 145 : audit [DBG] from='client.? 192.168.123.105:0/475018937' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:04.829590+0000 mon.a (mon.0) 145 : audit [DBG] from='client.? 192.168.123.105:0/475018937' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:05.802063+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:05.802063+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:06.565620+0000 mgr.a (mgr.14150) 32 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:06.565620+0000 mgr.a (mgr.14150) 32 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:06.801242+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:06.801242+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:06.802509+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:06.802509+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:07.802108+0000 mon.a (mon.0) 148 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:07.802108+0000 mon.a (mon.0) 148 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:08.565786+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:08.565786+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:08.802288+0000 mon.a (mon.0) 149 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:08.802288+0000 mon.a (mon.0) 149 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.802330+0000 mon.a (mon.0) 150 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.802330+0000 mon.a (mon.0) 150 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.815964+0000 mon.a (mon.0) 151 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.815964+0000 mon.a (mon.0) 151 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819666+0000 mon.a (mon.0) 152 : cluster [DBG] monmap epoch 2 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819666+0000 mon.a (mon.0) 152 : cluster [DBG] monmap epoch 2 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819717+0000 mon.a (mon.0) 153 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819717+0000 mon.a (mon.0) 153 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819758+0000 mon.a (mon.0) 154 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819758+0000 mon.a (mon.0) 154 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819798+0000 mon.a (mon.0) 155 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819798+0000 mon.a (mon.0) 155 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819838+0000 mon.a (mon.0) 156 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819838+0000 mon.a (mon.0) 156 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819879+0000 mon.a (mon.0) 157 : cluster [DBG] election_strategy: 1 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819879+0000 mon.a (mon.0) 157 : cluster [DBG] election_strategy: 1 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819919+0000 mon.a (mon.0) 158 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819919+0000 mon.a (mon.0) 158 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819964+0000 mon.a (mon.0) 159 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.819964+0000 mon.a (mon.0) 159 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.820361+0000 mon.a (mon.0) 160 : cluster [DBG] fsmap 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.820361+0000 mon.a (mon.0) 160 : cluster [DBG] fsmap 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.820456+0000 mon.a (mon.0) 161 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.820456+0000 mon.a (mon.0) 161 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.820685+0000 mon.a (mon.0) 162 : cluster [DBG] mgrmap e12: a(active, since 35s) 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.820685+0000 mon.a (mon.0) 162 : cluster [DBG] mgrmap e12: a(active, since 35s) 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.820961+0000 mon.a (mon.0) 163 : cluster [INF] overall HEALTH_OK 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: cluster 2026-03-08T23:35:09.820961+0000 mon.a (mon.0) 163 : cluster [INF] overall HEALTH_OK 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.826605+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.826605+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.831281+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.831281+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.832056+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.832056+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.832922+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:10.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:09 vm05 bash[22859]: audit 2026-03-08T23:35:09.832922+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.833739+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.833739+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.833852+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.833852+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.864290+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.864290+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.873680+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.873680+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.906757+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.906757+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.912065+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.912065+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.926080+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.926080+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.930811+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.930811+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.934989+0000 mon.a (mon.0) 172 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.934989+0000 mon.a (mon.0) 172 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.950089+0000 mon.a (mon.0) 173 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.950089+0000 mon.a (mon.0) 173 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.953799+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.953799+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.957209+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.957209+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.960166+0000 mon.a (mon.0) 176 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.960166+0000 mon.a (mon.0) 176 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.960488+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.960488+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.960762+0000 mon.a (mon.0) 177 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.960762+0000 mon.a (mon.0) 177 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.961156+0000 mon.a (mon.0) 178 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.961156+0000 mon.a (mon.0) 178 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.961489+0000 mon.a (mon.0) 179 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:09.961489+0000 mon.a (mon.0) 179 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.963326+0000 mgr.a (mgr.14150) 39 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-08T23:35:11.216 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:09.963326+0000 mgr.a (mgr.14150) 39 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.331245+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.331245+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.334892+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.334892+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:10.335471+0000 mgr.a (mgr.14150) 40 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:10.335471+0000 mgr.a (mgr.14150) 40 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.335695+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.335695+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.336133+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.336133+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.336564+0000 mon.a (mon.0) 184 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.336564+0000 mon.a (mon.0) 184 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:10.337144+0000 mgr.a (mgr.14150) 41 : cephadm [INF] Reconfiguring daemon mon.b on vm05 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: cephadm 2026-03-08T23:35:10.337144+0000 mgr.a (mgr.14150) 41 : cephadm [INF] Reconfiguring daemon mon.b on vm05 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.690235+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.690235+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.693289+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.693289+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.693932+0000 mon.a (mon.0) 187 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.693932+0000 mon.a (mon.0) 187 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.694747+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.694747+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.695127+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.695127+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.697759+0000 mon.a (mon.0) 190 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.697759+0000 mon.a (mon.0) 190 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.802439+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:11.217 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:10 vm02 bash[20612]: audit 2026-03-08T23:35:10.802439+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:11.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.833739+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-08T23:35:11.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.833739+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-08T23:35:11.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.833852+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.833852+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm05:/etc/ceph/ceph.conf 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.864290+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.864290+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.873680+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.873680+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/config/ceph.conf 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.906757+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.906757+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.912065+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.912065+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.926080+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.926080+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.930811+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.930811+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.934989+0000 mon.a (mon.0) 172 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.934989+0000 mon.a (mon.0) 172 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.950089+0000 mon.a (mon.0) 173 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.950089+0000 mon.a (mon.0) 173 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.953799+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.953799+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.957209+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.957209+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.960166+0000 mon.a (mon.0) 176 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.960166+0000 mon.a (mon.0) 176 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.960488+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.960488+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.960762+0000 mon.a (mon.0) 177 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.960762+0000 mon.a (mon.0) 177 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.961156+0000 mon.a (mon.0) 178 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.961156+0000 mon.a (mon.0) 178 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.961489+0000 mon.a (mon.0) 179 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:09.961489+0000 mon.a (mon.0) 179 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.963326+0000 mgr.a (mgr.14150) 39 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:09.963326+0000 mgr.a (mgr.14150) 39 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.331245+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.331245+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.334892+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.334892+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:10.335471+0000 mgr.a (mgr.14150) 40 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:10.335471+0000 mgr.a (mgr.14150) 40 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.335695+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.335695+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.336133+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.336133+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.336564+0000 mon.a (mon.0) 184 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.336564+0000 mon.a (mon.0) 184 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:10.337144+0000 mgr.a (mgr.14150) 41 : cephadm [INF] Reconfiguring daemon mon.b on vm05 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: cephadm 2026-03-08T23:35:10.337144+0000 mgr.a (mgr.14150) 41 : cephadm [INF] Reconfiguring daemon mon.b on vm05 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.690235+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.690235+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.693289+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.693289+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.693932+0000 mon.a (mon.0) 187 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:11.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.693932+0000 mon.a (mon.0) 187 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:11.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.694747+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.694747+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:11.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.695127+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:11.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.695127+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:11.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.697759+0000 mon.a (mon.0) 190 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.697759+0000 mon.a (mon.0) 190 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:11.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.802439+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:11.251 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:10 vm05 bash[22859]: audit 2026-03-08T23:35:10.802439+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:35:12.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:35:11 vm02 bash[20891]: debug 2026-03-08T23:35:11.800+0000 7f24db8d0640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-08T23:35:12.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:12 vm05 bash[22859]: cluster 2026-03-08T23:35:10.565931+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:12.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:12 vm05 bash[22859]: cluster 2026-03-08T23:35:10.565931+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:12.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:12 vm02 bash[20612]: cluster 2026-03-08T23:35:10.565931+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:12.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:12 vm02 bash[20612]: cluster 2026-03-08T23:35:10.565931+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:14.514 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:35:14.602 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:14 vm02 bash[20612]: cluster 2026-03-08T23:35:12.566120+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:14.602 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:14 vm02 bash[20612]: cluster 2026-03-08T23:35:12.566120+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:14.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:14 vm05 bash[22859]: cluster 2026-03-08T23:35:12.566120+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:14.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:14 vm05 bash[22859]: cluster 2026-03-08T23:35:12.566120+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:14.753 INFO:teuthology.orchestra.run.vm02.stdout:# minimal ceph.conf for 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:14.753 INFO:teuthology.orchestra.run.vm02.stdout:[global] 2026-03-08T23:35:14.753 INFO:teuthology.orchestra.run.vm02.stdout: fsid = 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:35:14.753 INFO:teuthology.orchestra.run.vm02.stdout: mon_host = [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] 2026-03-08T23:35:14.805 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-08T23:35:14.805 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:35:14.805 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.conf 2026-03-08T23:35:14.811 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:35:14.811 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:35:14.859 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:35:14.859 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.conf 2026-03-08T23:35:14.865 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:35:14.865 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-08T23:35:14.912 INFO:tasks.cephadm:Adding mgr.a on vm02 2026-03-08T23:35:14.912 INFO:tasks.cephadm:Adding mgr.b on vm05 2026-03-08T23:35:14.912 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch apply mgr '2;vm02=a;vm05=b' 2026-03-08T23:35:15.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:15 vm05 bash[22859]: audit 2026-03-08T23:35:14.755024+0000 mon.a (mon.0) 192 : audit [DBG] from='client.? 192.168.123.102:0/2661446072' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:15.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:15 vm05 bash[22859]: audit 2026-03-08T23:35:14.755024+0000 mon.a (mon.0) 192 : audit [DBG] from='client.? 192.168.123.102:0/2661446072' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:15.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:15 vm02 bash[20612]: audit 2026-03-08T23:35:14.755024+0000 mon.a (mon.0) 192 : audit [DBG] from='client.? 192.168.123.102:0/2661446072' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:15.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:15 vm02 bash[20612]: audit 2026-03-08T23:35:14.755024+0000 mon.a (mon.0) 192 : audit [DBG] from='client.? 192.168.123.102:0/2661446072' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:16.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:16 vm05 bash[22859]: cluster 2026-03-08T23:35:14.566336+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:16.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:16 vm05 bash[22859]: cluster 2026-03-08T23:35:14.566336+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:16.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:16 vm02 bash[20612]: cluster 2026-03-08T23:35:14.566336+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:16.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:16 vm02 bash[20612]: cluster 2026-03-08T23:35:14.566336+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:18.555 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.b/config 2026-03-08T23:35:18.608 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:18 vm05 bash[22859]: cluster 2026-03-08T23:35:16.566572+0000 mgr.a (mgr.14150) 45 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:18.608 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:18 vm05 bash[22859]: cluster 2026-03-08T23:35:16.566572+0000 mgr.a (mgr.14150) 45 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:18.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:18 vm02 bash[20612]: cluster 2026-03-08T23:35:16.566572+0000 mgr.a (mgr.14150) 45 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:18.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:18 vm02 bash[20612]: cluster 2026-03-08T23:35:16.566572+0000 mgr.a (mgr.14150) 45 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:18.825 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled mgr update... 2026-03-08T23:35:18.887 DEBUG:teuthology.orchestra.run.vm05:mgr.b> sudo journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.b.service 2026-03-08T23:35:18.888 INFO:tasks.cephadm:Deploying OSDs... 2026-03-08T23:35:18.888 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-08T23:35:18.888 DEBUG:teuthology.orchestra.run.vm02:> dd if=/scratch_devs of=/dev/stdout 2026-03-08T23:35:18.892 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:35:18.892 DEBUG:teuthology.orchestra.run.vm02:> ls /dev/[sv]d? 2026-03-08T23:35:18.935 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vda 2026-03-08T23:35:18.935 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdb 2026-03-08T23:35:18.935 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdc 2026-03-08T23:35:18.935 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdd 2026-03-08T23:35:18.935 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vde 2026-03-08T23:35:18.935 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-08T23:35:18.935 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-08T23:35:18.935 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdb 2026-03-08T23:35:18.979 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdb 2026-03-08T23:35:18.979 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-08T23:35:18.979 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-08T23:35:18.979 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-08T23:35:18.979 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-08 23:29:50.771395906 +0000 2026-03-08T23:35:18.979 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-08 23:29:49.867395906 +0000 2026-03-08T23:35:18.979 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-08 23:29:49.867395906 +0000 2026-03-08T23:35:18.979 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-08T23:35:18.979 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-08T23:35:19.027 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-08T23:35:19.027 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-08T23:35:19.027 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.00016581 s, 3.1 MB/s 2026-03-08T23:35:19.028 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-08T23:35:19.072 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdc 2026-03-08T23:35:19.119 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdc 2026-03-08T23:35:19.119 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-08T23:35:19.119 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-08T23:35:19.119 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-08T23:35:19.119 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-08 23:29:50.787395906 +0000 2026-03-08T23:35:19.119 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-08 23:29:49.827395906 +0000 2026-03-08T23:35:19.119 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-08 23:29:49.827395906 +0000 2026-03-08T23:35:19.119 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-08T23:35:19.120 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-08T23:35:19.144 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:19.168 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-08T23:35:19.168 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-08T23:35:19.168 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000166943 s, 3.1 MB/s 2026-03-08T23:35:19.168 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-08T23:35:19.217 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdd 2026-03-08T23:35:19.263 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdd 2026-03-08T23:35:19.264 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-08T23:35:19.264 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-08T23:35:19.264 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-08T23:35:19.264 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-08 23:29:50.771395906 +0000 2026-03-08T23:35:19.264 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-08 23:29:49.863395906 +0000 2026-03-08T23:35:19.264 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-08 23:29:49.863395906 +0000 2026-03-08T23:35:19.264 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-08T23:35:19.264 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-08T23:35:19.312 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-08T23:35:19.312 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-08T23:35:19.312 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000195355 s, 2.6 MB/s 2026-03-08T23:35:19.313 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-08T23:35:19.361 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vde 2026-03-08T23:35:19.408 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vde 2026-03-08T23:35:19.416 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-08T23:35:19.416 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-08T23:35:19.416 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-08T23:35:19.416 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-08 23:29:50.783395906 +0000 2026-03-08T23:35:19.416 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-08 23:29:49.871395906 +0000 2026-03-08T23:35:19.416 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-08 23:29:49.871395906 +0000 2026-03-08T23:35:19.416 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-08T23:35:19.416 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-08T23:35:19.455 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-08T23:35:19.455 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-08T23:35:19.455 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000190526 s, 2.7 MB/s 2026-03-08T23:35:19.456 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-08T23:35:19.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:19.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:19.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:19.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:19.501 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-08T23:35:19.501 DEBUG:teuthology.orchestra.run.vm05:> dd if=/scratch_devs of=/dev/stdout 2026-03-08T23:35:19.504 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:35:19.504 DEBUG:teuthology.orchestra.run.vm05:> ls /dev/[sv]d? 2026-03-08T23:35:19.549 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vda 2026-03-08T23:35:19.549 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdb 2026-03-08T23:35:19.549 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdc 2026-03-08T23:35:19.549 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdd 2026-03-08T23:35:19.549 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vde 2026-03-08T23:35:19.549 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-08T23:35:19.549 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-08T23:35:19.549 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdb 2026-03-08T23:35:19.593 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdb 2026-03-08T23:35:19.593 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-08T23:35:19.593 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-08T23:35:19.593 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-08T23:35:19.593 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-08 23:29:26.382044601 +0000 2026-03-08T23:35:19.593 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-08 23:29:25.294044601 +0000 2026-03-08T23:35:19.593 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-08 23:29:25.294044601 +0000 2026-03-08T23:35:19.593 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-08T23:35:19.593 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-08T23:35:19.655 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-08T23:35:19.655 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-08T23:35:19.655 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.00021365 s, 2.4 MB/s 2026-03-08T23:35:19.656 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-08T23:35:19.709 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdc 2026-03-08T23:35:19.764 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdc 2026-03-08T23:35:19.764 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-08T23:35:19.764 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-08T23:35:19.764 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-08T23:35:19.764 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-08 23:29:26.394044601 +0000 2026-03-08T23:35:19.764 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-08 23:29:25.290044601 +0000 2026-03-08T23:35:19.764 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-08 23:29:25.290044601 +0000 2026-03-08T23:35:19.764 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-08T23:35:19.764 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-08T23:35:19.786 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:19.786 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 systemd[1]: Started Ceph mgr.b for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:35:19.786 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:19.815 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-08T23:35:19.815 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-08T23:35:19.815 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000158867 s, 3.2 MB/s 2026-03-08T23:35:19.816 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-08T23:35:19.874 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdd 2026-03-08T23:35:19.920 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdd 2026-03-08T23:35:19.920 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-08T23:35:19.920 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-08T23:35:19.920 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-08T23:35:19.920 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-08 23:29:26.382044601 +0000 2026-03-08T23:35:19.921 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-08 23:29:25.290044601 +0000 2026-03-08T23:35:19.921 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-08 23:29:25.290044601 +0000 2026-03-08T23:35:19.921 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-08T23:35:19.921 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-08T23:35:19.957 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: cluster 2026-03-08T23:35:18.566795+0000 mgr.a (mgr.14150) 46 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:19.957 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: cluster 2026-03-08T23:35:18.566795+0000 mgr.a (mgr.14150) 46 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.821205+0000 mgr.a (mgr.14150) 47 : audit [DBG] from='client.14188 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm05=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.821205+0000 mgr.a (mgr.14150) 47 : audit [DBG] from='client.14188 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm05=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: cephadm 2026-03-08T23:35:18.822219+0000 mgr.a (mgr.14150) 48 : cephadm [INF] Saving service mgr spec with placement vm02=a;vm05=b;count:2 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: cephadm 2026-03-08T23:35:18.822219+0000 mgr.a (mgr.14150) 48 : cephadm [INF] Saving service mgr spec with placement vm02=a;vm05=b;count:2 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.825938+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.825938+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.826667+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.826667+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.827815+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.827815+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.828285+0000 mon.a (mon.0) 196 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.828285+0000 mon.a (mon.0) 196 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.832377+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.832377+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.833880+0000 mon.a (mon.0) 198 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.833880+0000 mon.a (mon.0) 198 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.835740+0000 mon.a (mon.0) 199 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.835740+0000 mon.a (mon.0) 199 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.837820+0000 mon.a (mon.0) 200 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.837820+0000 mon.a (mon.0) 200 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.838358+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:18.838358+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: cephadm 2026-03-08T23:35:18.838918+0000 mgr.a (mgr.14150) 49 : cephadm [INF] Deploying daemon mgr.b on vm05 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: cephadm 2026-03-08T23:35:18.838918+0000 mgr.a (mgr.14150) 49 : cephadm [INF] Deploying daemon mgr.b on vm05 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.600629+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.600629+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.604911+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.604911+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.608320+0000 mon.a (mon.0) 204 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.608320+0000 mon.a (mon.0) 204 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.611279+0000 mon.a (mon.0) 205 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.611279+0000 mon.a (mon.0) 205 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.621943+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[22859]: audit 2026-03-08T23:35:19.621943+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[23461]: debug 2026-03-08T23:35:19.777+0000 7f724dd51140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[23461]: debug 2026-03-08T23:35:19.817+0000 7f724dd51140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-08T23:35:19.958 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:19 vm05 bash[23461]: debug 2026-03-08T23:35:19.953+0000 7f724dd51140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-08T23:35:19.965 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-08T23:35:19.965 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-08T23:35:19.965 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000168064 s, 3.0 MB/s 2026-03-08T23:35:19.966 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-08T23:35:20.014 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vde 2026-03-08T23:35:20.060 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vde 2026-03-08T23:35:20.061 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-08T23:35:20.061 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-08T23:35:20.061 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-08T23:35:20.061 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-08 23:29:26.390044601 +0000 2026-03-08T23:35:20.061 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-08 23:29:25.294044601 +0000 2026-03-08T23:35:20.061 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-08 23:29:25.294044601 +0000 2026-03-08T23:35:20.061 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-08T23:35:20.061 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-08T23:35:20.108 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-08T23:35:20.108 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-08T23:35:20.108 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000164789 s, 3.1 MB/s 2026-03-08T23:35:20.109 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-08T23:35:20.154 INFO:tasks.cephadm:Deploying osd.0 on vm02 with /dev/vde... 2026-03-08T23:35:20.154 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- lvm zap /dev/vde 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: cluster 2026-03-08T23:35:18.566795+0000 mgr.a (mgr.14150) 46 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: cluster 2026-03-08T23:35:18.566795+0000 mgr.a (mgr.14150) 46 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.821205+0000 mgr.a (mgr.14150) 47 : audit [DBG] from='client.14188 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm05=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.821205+0000 mgr.a (mgr.14150) 47 : audit [DBG] from='client.14188 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm05=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: cephadm 2026-03-08T23:35:18.822219+0000 mgr.a (mgr.14150) 48 : cephadm [INF] Saving service mgr spec with placement vm02=a;vm05=b;count:2 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: cephadm 2026-03-08T23:35:18.822219+0000 mgr.a (mgr.14150) 48 : cephadm [INF] Saving service mgr spec with placement vm02=a;vm05=b;count:2 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.825938+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.825938+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.826667+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.826667+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.827815+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.827815+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.828285+0000 mon.a (mon.0) 196 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.828285+0000 mon.a (mon.0) 196 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.832377+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.832377+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.833880+0000 mon.a (mon.0) 198 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.833880+0000 mon.a (mon.0) 198 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.835740+0000 mon.a (mon.0) 199 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-08T23:35:20.161 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.835740+0000 mon.a (mon.0) 199 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.837820+0000 mon.a (mon.0) 200 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.837820+0000 mon.a (mon.0) 200 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.838358+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:18.838358+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: cephadm 2026-03-08T23:35:18.838918+0000 mgr.a (mgr.14150) 49 : cephadm [INF] Deploying daemon mgr.b on vm05 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: cephadm 2026-03-08T23:35:18.838918+0000 mgr.a (mgr.14150) 49 : cephadm [INF] Deploying daemon mgr.b on vm05 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.600629+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.600629+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.604911+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.604911+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.608320+0000 mon.a (mon.0) 204 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.608320+0000 mon.a (mon.0) 204 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.611279+0000 mon.a (mon.0) 205 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.611279+0000 mon.a (mon.0) 205 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.621943+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:20.162 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:19 vm02 bash[20612]: audit 2026-03-08T23:35:19.621943+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:20.694 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:20 vm05 bash[23461]: debug 2026-03-08T23:35:20.253+0000 7f724dd51140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-08T23:35:20.999 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:20 vm05 bash[23461]: debug 2026-03-08T23:35:20.689+0000 7f724dd51140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-08T23:35:20.999 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:20 vm05 bash[23461]: debug 2026-03-08T23:35:20.773+0000 7f724dd51140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-08T23:35:20.999 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:20 vm05 bash[23461]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-08T23:35:20.999 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:20 vm05 bash[23461]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-08T23:35:20.999 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:20 vm05 bash[23461]: from numpy import show_config as show_numpy_config 2026-03-08T23:35:20.999 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:20 vm05 bash[23461]: debug 2026-03-08T23:35:20.893+0000 7f724dd51140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-08T23:35:21.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.037+0000 7f724dd51140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-08T23:35:21.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.073+0000 7f724dd51140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-08T23:35:21.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.113+0000 7f724dd51140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-08T23:35:21.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.157+0000 7f724dd51140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-08T23:35:21.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.217+0000 7f724dd51140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-08T23:35:21.941 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[22859]: cluster 2026-03-08T23:35:20.566996+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:21.941 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[22859]: cluster 2026-03-08T23:35:20.566996+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:21.941 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.653+0000 7f724dd51140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-08T23:35:21.941 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.689+0000 7f724dd51140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-08T23:35:21.941 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.725+0000 7f724dd51140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-08T23:35:21.941 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.861+0000 7f724dd51140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-08T23:35:21.941 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.901+0000 7f724dd51140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-08T23:35:22.197 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:21 vm05 bash[23461]: debug 2026-03-08T23:35:21.937+0000 7f724dd51140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-08T23:35:22.197 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[23461]: debug 2026-03-08T23:35:22.045+0000 7f724dd51140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-08T23:35:22.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:21 vm02 bash[20612]: cluster 2026-03-08T23:35:20.566996+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:22.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:21 vm02 bash[20612]: cluster 2026-03-08T23:35:20.566996+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:22.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[23461]: debug 2026-03-08T23:35:22.193+0000 7f724dd51140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-08T23:35:22.499 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[23461]: debug 2026-03-08T23:35:22.361+0000 7f724dd51140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-08T23:35:22.500 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[23461]: debug 2026-03-08T23:35:22.397+0000 7f724dd51140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-08T23:35:22.500 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[23461]: debug 2026-03-08T23:35:22.441+0000 7f724dd51140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-08T23:35:22.999 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[23461]: debug 2026-03-08T23:35:22.589+0000 7f724dd51140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-08T23:35:22.999 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[23461]: debug 2026-03-08T23:35:22.809+0000 7f724dd51140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-08T23:35:22.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: audit 2026-03-08T23:35:22.815606+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-08T23:35:22.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: audit 2026-03-08T23:35:22.815606+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-08T23:35:22.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: audit 2026-03-08T23:35:22.816440+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-08T23:35:22.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: audit 2026-03-08T23:35:22.816440+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-08T23:35:22.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: cluster 2026-03-08T23:35:22.817080+0000 mon.a (mon.0) 207 : cluster [DBG] Standby manager daemon b started 2026-03-08T23:35:23.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: cluster 2026-03-08T23:35:22.817080+0000 mon.a (mon.0) 207 : cluster [DBG] Standby manager daemon b started 2026-03-08T23:35:23.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: audit 2026-03-08T23:35:22.817274+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-08T23:35:23.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: audit 2026-03-08T23:35:22.817274+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-08T23:35:23.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: audit 2026-03-08T23:35:22.817897+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-08T23:35:23.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:22 vm05 bash[22859]: audit 2026-03-08T23:35:22.817897+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: audit 2026-03-08T23:35:22.815606+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: audit 2026-03-08T23:35:22.815606+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: audit 2026-03-08T23:35:22.816440+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: audit 2026-03-08T23:35:22.816440+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: cluster 2026-03-08T23:35:22.817080+0000 mon.a (mon.0) 207 : cluster [DBG] Standby manager daemon b started 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: cluster 2026-03-08T23:35:22.817080+0000 mon.a (mon.0) 207 : cluster [DBG] Standby manager daemon b started 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: audit 2026-03-08T23:35:22.817274+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: audit 2026-03-08T23:35:22.817274+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: audit 2026-03-08T23:35:22.817897+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-08T23:35:23.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:22 vm02 bash[20612]: audit 2026-03-08T23:35:22.817897+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.105:0/714904968' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-08T23:35:24.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:23 vm05 bash[22859]: cluster 2026-03-08T23:35:22.567230+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:24.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:23 vm05 bash[22859]: cluster 2026-03-08T23:35:22.567230+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:24.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:23 vm05 bash[22859]: cluster 2026-03-08T23:35:22.853196+0000 mon.a (mon.0) 208 : cluster [DBG] mgrmap e13: a(active, since 48s), standbys: b 2026-03-08T23:35:24.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:23 vm05 bash[22859]: cluster 2026-03-08T23:35:22.853196+0000 mon.a (mon.0) 208 : cluster [DBG] mgrmap e13: a(active, since 48s), standbys: b 2026-03-08T23:35:24.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:23 vm05 bash[22859]: audit 2026-03-08T23:35:22.853271+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-08T23:35:24.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:23 vm05 bash[22859]: audit 2026-03-08T23:35:22.853271+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-08T23:35:24.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:23 vm02 bash[20612]: cluster 2026-03-08T23:35:22.567230+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:24.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:23 vm02 bash[20612]: cluster 2026-03-08T23:35:22.567230+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:24.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:23 vm02 bash[20612]: cluster 2026-03-08T23:35:22.853196+0000 mon.a (mon.0) 208 : cluster [DBG] mgrmap e13: a(active, since 48s), standbys: b 2026-03-08T23:35:24.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:23 vm02 bash[20612]: cluster 2026-03-08T23:35:22.853196+0000 mon.a (mon.0) 208 : cluster [DBG] mgrmap e13: a(active, since 48s), standbys: b 2026-03-08T23:35:24.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:23 vm02 bash[20612]: audit 2026-03-08T23:35:22.853271+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-08T23:35:24.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:23 vm02 bash[20612]: audit 2026-03-08T23:35:22.853271+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-08T23:35:24.765 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:35:25.748 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:35:25.762 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch daemon add osd vm02:/dev/vde 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.557533+0000 mon.a (mon.0) 210 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.557533+0000 mon.a (mon.0) 210 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.562170+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.562170+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.563904+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.563904+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.564365+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.564365+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: cluster 2026-03-08T23:35:24.567392+0000 mgr.a (mgr.14150) 52 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: cluster 2026-03-08T23:35:24.567392+0000 mgr.a (mgr.14150) 52 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.568341+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.568341+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: cephadm 2026-03-08T23:35:24.578292+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: cephadm 2026-03-08T23:35:24.578292+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.578553+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.578553+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-08T23:35:25.813 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.579160+0000 mon.a (mon.0) 216 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.579160+0000 mon.a (mon.0) 216 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.579547+0000 mon.a (mon.0) 217 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.579547+0000 mon.a (mon.0) 217 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: cephadm 2026-03-08T23:35:24.579938+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Reconfiguring daemon mgr.a on vm02 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: cephadm 2026-03-08T23:35:24.579938+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Reconfiguring daemon mgr.a on vm02 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.598440+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:24.598440+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.007157+0000 mon.a (mon.0) 219 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.007157+0000 mon.a (mon.0) 219 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.011487+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.011487+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.012874+0000 mon.a (mon.0) 221 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.012874+0000 mon.a (mon.0) 221 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.013786+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.013786+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.014129+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.014129+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.018738+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.814 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:25 vm02 bash[20612]: audit 2026-03-08T23:35:25.018738+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.557533+0000 mon.a (mon.0) 210 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.557533+0000 mon.a (mon.0) 210 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.562170+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.562170+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:25.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.563904+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:25.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.563904+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:25.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.564365+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.564365+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: cluster 2026-03-08T23:35:24.567392+0000 mgr.a (mgr.14150) 52 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: cluster 2026-03-08T23:35:24.567392+0000 mgr.a (mgr.14150) 52 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.568341+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.568341+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: cephadm 2026-03-08T23:35:24.578292+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: cephadm 2026-03-08T23:35:24.578292+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.578553+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.578553+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.579160+0000 mon.a (mon.0) 216 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.579160+0000 mon.a (mon.0) 216 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.579547+0000 mon.a (mon.0) 217 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.579547+0000 mon.a (mon.0) 217 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: cephadm 2026-03-08T23:35:24.579938+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Reconfiguring daemon mgr.a on vm02 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: cephadm 2026-03-08T23:35:24.579938+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Reconfiguring daemon mgr.a on vm02 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.598440+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:24.598440+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.007157+0000 mon.a (mon.0) 219 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.007157+0000 mon.a (mon.0) 219 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.011487+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.011487+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.012874+0000 mon.a (mon.0) 221 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.012874+0000 mon.a (mon.0) 221 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.013786+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.013786+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.014129+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.014129+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.018738+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:26.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:25 vm05 bash[22859]: audit 2026-03-08T23:35:25.018738+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:27.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:27 vm05 bash[22859]: cluster 2026-03-08T23:35:26.567565+0000 mgr.a (mgr.14150) 55 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:27.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:27 vm05 bash[22859]: cluster 2026-03-08T23:35:26.567565+0000 mgr.a (mgr.14150) 55 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:28.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:27 vm02 bash[20612]: cluster 2026-03-08T23:35:26.567565+0000 mgr.a (mgr.14150) 55 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:28.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:27 vm02 bash[20612]: cluster 2026-03-08T23:35:26.567565+0000 mgr.a (mgr.14150) 55 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:29.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:29 vm05 bash[22859]: cluster 2026-03-08T23:35:28.567770+0000 mgr.a (mgr.14150) 56 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:29.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:29 vm05 bash[22859]: cluster 2026-03-08T23:35:28.567770+0000 mgr.a (mgr.14150) 56 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:30.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:29 vm02 bash[20612]: cluster 2026-03-08T23:35:28.567770+0000 mgr.a (mgr.14150) 56 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:30.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:29 vm02 bash[20612]: cluster 2026-03-08T23:35:28.567770+0000 mgr.a (mgr.14150) 56 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:30.433 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:35:31.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: cluster 2026-03-08T23:35:30.567965+0000 mgr.a (mgr.14150) 57 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: cluster 2026-03-08T23:35:30.567965+0000 mgr.a (mgr.14150) 57 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: audit 2026-03-08T23:35:30.700016+0000 mgr.a (mgr.14150) 58 : audit [DBG] from='client.14194 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: audit 2026-03-08T23:35:30.700016+0000 mgr.a (mgr.14150) 58 : audit [DBG] from='client.14194 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: audit 2026-03-08T23:35:30.701685+0000 mon.a (mon.0) 225 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: audit 2026-03-08T23:35:30.701685+0000 mon.a (mon.0) 225 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: audit 2026-03-08T23:35:30.702991+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: audit 2026-03-08T23:35:30.702991+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: audit 2026-03-08T23:35:30.703382+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:32.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:31 vm05 bash[22859]: audit 2026-03-08T23:35:30.703382+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:32.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: cluster 2026-03-08T23:35:30.567965+0000 mgr.a (mgr.14150) 57 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:32.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: cluster 2026-03-08T23:35:30.567965+0000 mgr.a (mgr.14150) 57 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:32.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: audit 2026-03-08T23:35:30.700016+0000 mgr.a (mgr.14150) 58 : audit [DBG] from='client.14194 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:32.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: audit 2026-03-08T23:35:30.700016+0000 mgr.a (mgr.14150) 58 : audit [DBG] from='client.14194 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:35:32.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: audit 2026-03-08T23:35:30.701685+0000 mon.a (mon.0) 225 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-08T23:35:32.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: audit 2026-03-08T23:35:30.701685+0000 mon.a (mon.0) 225 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-08T23:35:32.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: audit 2026-03-08T23:35:30.702991+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-08T23:35:32.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: audit 2026-03-08T23:35:30.702991+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-08T23:35:32.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: audit 2026-03-08T23:35:30.703382+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:32.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:31 vm02 bash[20612]: audit 2026-03-08T23:35:30.703382+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:33.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:33 vm05 bash[22859]: cluster 2026-03-08T23:35:32.568139+0000 mgr.a (mgr.14150) 59 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:33.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:33 vm05 bash[22859]: cluster 2026-03-08T23:35:32.568139+0000 mgr.a (mgr.14150) 59 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:34.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:33 vm02 bash[20612]: cluster 2026-03-08T23:35:32.568139+0000 mgr.a (mgr.14150) 59 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:34.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:33 vm02 bash[20612]: cluster 2026-03-08T23:35:32.568139+0000 mgr.a (mgr.14150) 59 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:35.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:35 vm02 bash[20612]: cluster 2026-03-08T23:35:34.568321+0000 mgr.a (mgr.14150) 60 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:35.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:35 vm02 bash[20612]: cluster 2026-03-08T23:35:34.568321+0000 mgr.a (mgr.14150) 60 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:35.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:35 vm05 bash[22859]: cluster 2026-03-08T23:35:34.568321+0000 mgr.a (mgr.14150) 60 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:35.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:35 vm05 bash[22859]: cluster 2026-03-08T23:35:34.568321+0000 mgr.a (mgr.14150) 60 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:36.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:36 vm02 bash[20612]: audit 2026-03-08T23:35:36.043729+0000 mon.a (mon.0) 228 : audit [INF] from='client.? 192.168.123.102:0/781000070' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6663b6e9-7873-428d-8f01-0747ce138477"}]: dispatch 2026-03-08T23:35:36.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:36 vm02 bash[20612]: audit 2026-03-08T23:35:36.043729+0000 mon.a (mon.0) 228 : audit [INF] from='client.? 192.168.123.102:0/781000070' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6663b6e9-7873-428d-8f01-0747ce138477"}]: dispatch 2026-03-08T23:35:36.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:36 vm02 bash[20612]: audit 2026-03-08T23:35:36.047643+0000 mon.a (mon.0) 229 : audit [INF] from='client.? 192.168.123.102:0/781000070' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6663b6e9-7873-428d-8f01-0747ce138477"}]': finished 2026-03-08T23:35:36.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:36 vm02 bash[20612]: audit 2026-03-08T23:35:36.047643+0000 mon.a (mon.0) 229 : audit [INF] from='client.? 192.168.123.102:0/781000070' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6663b6e9-7873-428d-8f01-0747ce138477"}]': finished 2026-03-08T23:35:36.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:36 vm02 bash[20612]: cluster 2026-03-08T23:35:36.051426+0000 mon.a (mon.0) 230 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-08T23:35:36.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:36 vm02 bash[20612]: cluster 2026-03-08T23:35:36.051426+0000 mon.a (mon.0) 230 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-08T23:35:36.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:36 vm02 bash[20612]: audit 2026-03-08T23:35:36.051689+0000 mon.a (mon.0) 231 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:36.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:36 vm02 bash[20612]: audit 2026-03-08T23:35:36.051689+0000 mon.a (mon.0) 231 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:36.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:36 vm05 bash[22859]: audit 2026-03-08T23:35:36.043729+0000 mon.a (mon.0) 228 : audit [INF] from='client.? 192.168.123.102:0/781000070' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6663b6e9-7873-428d-8f01-0747ce138477"}]: dispatch 2026-03-08T23:35:37.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:36 vm05 bash[22859]: audit 2026-03-08T23:35:36.043729+0000 mon.a (mon.0) 228 : audit [INF] from='client.? 192.168.123.102:0/781000070' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6663b6e9-7873-428d-8f01-0747ce138477"}]: dispatch 2026-03-08T23:35:37.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:36 vm05 bash[22859]: audit 2026-03-08T23:35:36.047643+0000 mon.a (mon.0) 229 : audit [INF] from='client.? 192.168.123.102:0/781000070' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6663b6e9-7873-428d-8f01-0747ce138477"}]': finished 2026-03-08T23:35:37.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:36 vm05 bash[22859]: audit 2026-03-08T23:35:36.047643+0000 mon.a (mon.0) 229 : audit [INF] from='client.? 192.168.123.102:0/781000070' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6663b6e9-7873-428d-8f01-0747ce138477"}]': finished 2026-03-08T23:35:37.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:36 vm05 bash[22859]: cluster 2026-03-08T23:35:36.051426+0000 mon.a (mon.0) 230 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-08T23:35:37.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:36 vm05 bash[22859]: cluster 2026-03-08T23:35:36.051426+0000 mon.a (mon.0) 230 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-08T23:35:37.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:36 vm05 bash[22859]: audit 2026-03-08T23:35:36.051689+0000 mon.a (mon.0) 231 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:37.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:36 vm05 bash[22859]: audit 2026-03-08T23:35:36.051689+0000 mon.a (mon.0) 231 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:37.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:37 vm05 bash[22859]: cluster 2026-03-08T23:35:36.568536+0000 mgr.a (mgr.14150) 61 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:38.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:37 vm05 bash[22859]: cluster 2026-03-08T23:35:36.568536+0000 mgr.a (mgr.14150) 61 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:38.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:37 vm05 bash[22859]: audit 2026-03-08T23:35:36.642723+0000 mon.a (mon.0) 232 : audit [DBG] from='client.? 192.168.123.102:0/1395998512' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-08T23:35:38.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:37 vm05 bash[22859]: audit 2026-03-08T23:35:36.642723+0000 mon.a (mon.0) 232 : audit [DBG] from='client.? 192.168.123.102:0/1395998512' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-08T23:35:38.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:37 vm02 bash[20612]: cluster 2026-03-08T23:35:36.568536+0000 mgr.a (mgr.14150) 61 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:38.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:37 vm02 bash[20612]: cluster 2026-03-08T23:35:36.568536+0000 mgr.a (mgr.14150) 61 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:38.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:37 vm02 bash[20612]: audit 2026-03-08T23:35:36.642723+0000 mon.a (mon.0) 232 : audit [DBG] from='client.? 192.168.123.102:0/1395998512' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-08T23:35:38.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:37 vm02 bash[20612]: audit 2026-03-08T23:35:36.642723+0000 mon.a (mon.0) 232 : audit [DBG] from='client.? 192.168.123.102:0/1395998512' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-08T23:35:39.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:39 vm05 bash[22859]: cluster 2026-03-08T23:35:38.568753+0000 mgr.a (mgr.14150) 62 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:40.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:39 vm05 bash[22859]: cluster 2026-03-08T23:35:38.568753+0000 mgr.a (mgr.14150) 62 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:40.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:39 vm02 bash[20612]: cluster 2026-03-08T23:35:38.568753+0000 mgr.a (mgr.14150) 62 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:40.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:39 vm02 bash[20612]: cluster 2026-03-08T23:35:38.568753+0000 mgr.a (mgr.14150) 62 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:42.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:41 vm02 bash[20612]: cluster 2026-03-08T23:35:40.568953+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:42.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:41 vm02 bash[20612]: cluster 2026-03-08T23:35:40.568953+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:42.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:41 vm05 bash[22859]: cluster 2026-03-08T23:35:40.568953+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:42.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:41 vm05 bash[22859]: cluster 2026-03-08T23:35:40.568953+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:44.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:43 vm02 bash[20612]: cluster 2026-03-08T23:35:42.569121+0000 mgr.a (mgr.14150) 64 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:44.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:43 vm02 bash[20612]: cluster 2026-03-08T23:35:42.569121+0000 mgr.a (mgr.14150) 64 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:44.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:43 vm05 bash[22859]: cluster 2026-03-08T23:35:42.569121+0000 mgr.a (mgr.14150) 64 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:44.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:43 vm05 bash[22859]: cluster 2026-03-08T23:35:42.569121+0000 mgr.a (mgr.14150) 64 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:45.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:45.608 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:35:45 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:45.878 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:35:45 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 bash[20612]: cluster 2026-03-08T23:35:44.569304+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 bash[20612]: cluster 2026-03-08T23:35:44.569304+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 bash[20612]: audit 2026-03-08T23:35:44.820061+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 bash[20612]: audit 2026-03-08T23:35:44.820061+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 bash[20612]: audit 2026-03-08T23:35:44.820698+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 bash[20612]: audit 2026-03-08T23:35:44.820698+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 bash[20612]: cephadm 2026-03-08T23:35:44.821175+0000 mgr.a (mgr.14150) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-08T23:35:45.879 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:45 vm02 bash[20612]: cephadm 2026-03-08T23:35:44.821175+0000 mgr.a (mgr.14150) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-08T23:35:46.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:45 vm05 bash[22859]: cluster 2026-03-08T23:35:44.569304+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:46.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:45 vm05 bash[22859]: cluster 2026-03-08T23:35:44.569304+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:46.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:45 vm05 bash[22859]: audit 2026-03-08T23:35:44.820061+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-08T23:35:46.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:45 vm05 bash[22859]: audit 2026-03-08T23:35:44.820061+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-08T23:35:46.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:45 vm05 bash[22859]: audit 2026-03-08T23:35:44.820698+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:46.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:45 vm05 bash[22859]: audit 2026-03-08T23:35:44.820698+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:46.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:45 vm05 bash[22859]: cephadm 2026-03-08T23:35:44.821175+0000 mgr.a (mgr.14150) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-08T23:35:46.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:45 vm05 bash[22859]: cephadm 2026-03-08T23:35:44.821175+0000 mgr.a (mgr.14150) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-08T23:35:47.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:46 vm05 bash[22859]: audit 2026-03-08T23:35:45.821828+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:47.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:46 vm05 bash[22859]: audit 2026-03-08T23:35:45.821828+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:47.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:46 vm05 bash[22859]: audit 2026-03-08T23:35:45.826146+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:47.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:46 vm05 bash[22859]: audit 2026-03-08T23:35:45.826146+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:47.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:46 vm05 bash[22859]: audit 2026-03-08T23:35:45.830798+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:47.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:46 vm05 bash[22859]: audit 2026-03-08T23:35:45.830798+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:47.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:46 vm02 bash[20612]: audit 2026-03-08T23:35:45.821828+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:47.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:46 vm02 bash[20612]: audit 2026-03-08T23:35:45.821828+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:47.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:46 vm02 bash[20612]: audit 2026-03-08T23:35:45.826146+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:47.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:46 vm02 bash[20612]: audit 2026-03-08T23:35:45.826146+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:47.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:46 vm02 bash[20612]: audit 2026-03-08T23:35:45.830798+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:47.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:46 vm02 bash[20612]: audit 2026-03-08T23:35:45.830798+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:48.184 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:47 vm02 bash[20612]: cluster 2026-03-08T23:35:46.569473+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:48.184 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:47 vm02 bash[20612]: cluster 2026-03-08T23:35:46.569473+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:48.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:47 vm05 bash[22859]: cluster 2026-03-08T23:35:46.569473+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:48.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:47 vm05 bash[22859]: cluster 2026-03-08T23:35:46.569473+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:50.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:49 vm05 bash[22859]: cluster 2026-03-08T23:35:48.569697+0000 mgr.a (mgr.14150) 68 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:50.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:49 vm05 bash[22859]: cluster 2026-03-08T23:35:48.569697+0000 mgr.a (mgr.14150) 68 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:50.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:49 vm05 bash[22859]: audit 2026-03-08T23:35:49.132880+0000 mon.a (mon.0) 238 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-08T23:35:50.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:49 vm05 bash[22859]: audit 2026-03-08T23:35:49.132880+0000 mon.a (mon.0) 238 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-08T23:35:50.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:49 vm02 bash[20612]: cluster 2026-03-08T23:35:48.569697+0000 mgr.a (mgr.14150) 68 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:50.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:49 vm02 bash[20612]: cluster 2026-03-08T23:35:48.569697+0000 mgr.a (mgr.14150) 68 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:50.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:49 vm02 bash[20612]: audit 2026-03-08T23:35:49.132880+0000 mon.a (mon.0) 238 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-08T23:35:50.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:49 vm02 bash[20612]: audit 2026-03-08T23:35:49.132880+0000 mon.a (mon.0) 238 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-08T23:35:51.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:50 vm05 bash[22859]: audit 2026-03-08T23:35:49.844664+0000 mon.a (mon.0) 239 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-08T23:35:51.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:50 vm05 bash[22859]: audit 2026-03-08T23:35:49.844664+0000 mon.a (mon.0) 239 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-08T23:35:51.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:50 vm05 bash[22859]: cluster 2026-03-08T23:35:49.846537+0000 mon.a (mon.0) 240 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-08T23:35:51.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:50 vm05 bash[22859]: cluster 2026-03-08T23:35:49.846537+0000 mon.a (mon.0) 240 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-08T23:35:51.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:50 vm05 bash[22859]: audit 2026-03-08T23:35:49.846707+0000 mon.a (mon.0) 241 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-08T23:35:51.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:50 vm05 bash[22859]: audit 2026-03-08T23:35:49.846707+0000 mon.a (mon.0) 241 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-08T23:35:51.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:50 vm05 bash[22859]: audit 2026-03-08T23:35:49.846810+0000 mon.a (mon.0) 242 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:51.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:50 vm05 bash[22859]: audit 2026-03-08T23:35:49.846810+0000 mon.a (mon.0) 242 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:51.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:50 vm02 bash[20612]: audit 2026-03-08T23:35:49.844664+0000 mon.a (mon.0) 239 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-08T23:35:51.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:50 vm02 bash[20612]: audit 2026-03-08T23:35:49.844664+0000 mon.a (mon.0) 239 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-08T23:35:51.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:50 vm02 bash[20612]: cluster 2026-03-08T23:35:49.846537+0000 mon.a (mon.0) 240 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-08T23:35:51.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:50 vm02 bash[20612]: cluster 2026-03-08T23:35:49.846537+0000 mon.a (mon.0) 240 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-08T23:35:51.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:50 vm02 bash[20612]: audit 2026-03-08T23:35:49.846707+0000 mon.a (mon.0) 241 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-08T23:35:51.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:50 vm02 bash[20612]: audit 2026-03-08T23:35:49.846707+0000 mon.a (mon.0) 241 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-08T23:35:51.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:50 vm02 bash[20612]: audit 2026-03-08T23:35:49.846810+0000 mon.a (mon.0) 242 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:51.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:50 vm02 bash[20612]: audit 2026-03-08T23:35:49.846810+0000 mon.a (mon.0) 242 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: cluster 2026-03-08T23:35:50.569855+0000 mgr.a (mgr.14150) 69 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:52.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: cluster 2026-03-08T23:35:50.569855+0000 mgr.a (mgr.14150) 69 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:52.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: audit 2026-03-08T23:35:50.849585+0000 mon.a (mon.0) 243 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: audit 2026-03-08T23:35:50.849585+0000 mon.a (mon.0) 243 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: cluster 2026-03-08T23:35:50.851865+0000 mon.a (mon.0) 244 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: cluster 2026-03-08T23:35:50.851865+0000 mon.a (mon.0) 244 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: audit 2026-03-08T23:35:50.852815+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: audit 2026-03-08T23:35:50.852815+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: audit 2026-03-08T23:35:50.854907+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: audit 2026-03-08T23:35:50.854907+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: audit 2026-03-08T23:35:51.854917+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:51 vm02 bash[20612]: audit 2026-03-08T23:35:51.854917+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: cluster 2026-03-08T23:35:50.569855+0000 mgr.a (mgr.14150) 69 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: cluster 2026-03-08T23:35:50.569855+0000 mgr.a (mgr.14150) 69 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: audit 2026-03-08T23:35:50.849585+0000 mon.a (mon.0) 243 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: audit 2026-03-08T23:35:50.849585+0000 mon.a (mon.0) 243 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: cluster 2026-03-08T23:35:50.851865+0000 mon.a (mon.0) 244 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: cluster 2026-03-08T23:35:50.851865+0000 mon.a (mon.0) 244 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: audit 2026-03-08T23:35:50.852815+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: audit 2026-03-08T23:35:50.852815+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: audit 2026-03-08T23:35:50.854907+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: audit 2026-03-08T23:35:50.854907+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: audit 2026-03-08T23:35:51.854917+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:52.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:51 vm05 bash[22859]: audit 2026-03-08T23:35:51.854917+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:53.018 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 0 on host 'vm02' 2026-03-08T23:35:53.031 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: cluster 2026-03-08T23:35:50.129541+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: cluster 2026-03-08T23:35:50.129541+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: cluster 2026-03-08T23:35:50.129594+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: cluster 2026-03-08T23:35:50.129594+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: cluster 2026-03-08T23:35:51.860410+0000 mon.a (mon.0) 248 : cluster [INF] osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430] boot 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: cluster 2026-03-08T23:35:51.860410+0000 mon.a (mon.0) 248 : cluster [INF] osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430] boot 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: cluster 2026-03-08T23:35:51.860431+0000 mon.a (mon.0) 249 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: cluster 2026-03-08T23:35:51.860431+0000 mon.a (mon.0) 249 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:51.861070+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:51.861070+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:51.997263+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:51.997263+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:52.006117+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:52.006117+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:52.419165+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:52.419165+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:52.419808+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:52.419808+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:52.433998+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.032 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:52 vm02 bash[20612]: audit 2026-03-08T23:35:52.433998+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.098 DEBUG:teuthology.orchestra.run.vm02:osd.0> sudo journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@osd.0.service 2026-03-08T23:35:53.141 INFO:tasks.cephadm:Deploying osd.1 on vm05 with /dev/vde... 2026-03-08T23:35:53.142 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- lvm zap /dev/vde 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: cluster 2026-03-08T23:35:50.129541+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: cluster 2026-03-08T23:35:50.129541+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: cluster 2026-03-08T23:35:50.129594+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: cluster 2026-03-08T23:35:50.129594+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: cluster 2026-03-08T23:35:51.860410+0000 mon.a (mon.0) 248 : cluster [INF] osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430] boot 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: cluster 2026-03-08T23:35:51.860410+0000 mon.a (mon.0) 248 : cluster [INF] osd.0 [v2:192.168.123.102:6802/1666513430,v1:192.168.123.102:6803/1666513430] boot 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: cluster 2026-03-08T23:35:51.860431+0000 mon.a (mon.0) 249 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: cluster 2026-03-08T23:35:51.860431+0000 mon.a (mon.0) 249 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:51.861070+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:51.861070+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:51.997263+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:51.997263+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:52.006117+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:52.006117+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:52.419165+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:52.419165+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:52.419808+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:52.419808+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:52.433998+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:53.149 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:52 vm05 bash[22859]: audit 2026-03-08T23:35:52.433998+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:54.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: cluster 2026-03-08T23:35:52.570029+0000 mgr.a (mgr.14150) 70 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: cluster 2026-03-08T23:35:52.570029+0000 mgr.a (mgr.14150) 70 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: audit 2026-03-08T23:35:53.002948+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: audit 2026-03-08T23:35:53.002948+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: audit 2026-03-08T23:35:53.008024+0000 mon.a (mon.0) 257 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: audit 2026-03-08T23:35:53.008024+0000 mon.a (mon.0) 257 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: cluster 2026-03-08T23:35:53.012779+0000 mon.a (mon.0) 258 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: cluster 2026-03-08T23:35:53.012779+0000 mon.a (mon.0) 258 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: audit 2026-03-08T23:35:53.015741+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:54.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:53 vm05 bash[22859]: audit 2026-03-08T23:35:53.015741+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:54.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: cluster 2026-03-08T23:35:52.570029+0000 mgr.a (mgr.14150) 70 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:54.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: cluster 2026-03-08T23:35:52.570029+0000 mgr.a (mgr.14150) 70 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-08T23:35:54.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: audit 2026-03-08T23:35:53.002948+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:54.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: audit 2026-03-08T23:35:53.002948+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:35:54.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: audit 2026-03-08T23:35:53.008024+0000 mon.a (mon.0) 257 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:54.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: audit 2026-03-08T23:35:53.008024+0000 mon.a (mon.0) 257 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:54.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: cluster 2026-03-08T23:35:53.012779+0000 mon.a (mon.0) 258 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-08T23:35:54.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: cluster 2026-03-08T23:35:53.012779+0000 mon.a (mon.0) 258 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-08T23:35:54.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: audit 2026-03-08T23:35:53.015741+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:54.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:53 vm02 bash[20612]: audit 2026-03-08T23:35:53.015741+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:35:56.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:55 vm05 bash[22859]: cluster 2026-03-08T23:35:54.570226+0000 mgr.a (mgr.14150) 71 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:35:56.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:55 vm05 bash[22859]: cluster 2026-03-08T23:35:54.570226+0000 mgr.a (mgr.14150) 71 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:35:56.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:55 vm02 bash[20612]: cluster 2026-03-08T23:35:54.570226+0000 mgr.a (mgr.14150) 71 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:35:56.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:55 vm02 bash[20612]: cluster 2026-03-08T23:35:54.570226+0000 mgr.a (mgr.14150) 71 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:35:57.756 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.b/config 2026-03-08T23:35:58.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:57 vm05 bash[22859]: cluster 2026-03-08T23:35:56.570430+0000 mgr.a (mgr.14150) 72 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:35:58.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:57 vm05 bash[22859]: cluster 2026-03-08T23:35:56.570430+0000 mgr.a (mgr.14150) 72 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:35:58.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:57 vm02 bash[20612]: cluster 2026-03-08T23:35:56.570430+0000 mgr.a (mgr.14150) 72 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:35:58.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:57 vm02 bash[20612]: cluster 2026-03-08T23:35:56.570430+0000 mgr.a (mgr.14150) 72 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:35:58.889 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-08T23:35:58.903 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch daemon add osd vm05:/dev/vde 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: cluster 2026-03-08T23:35:58.570674+0000 mgr.a (mgr.14150) 73 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: cluster 2026-03-08T23:35:58.570674+0000 mgr.a (mgr.14150) 73 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: cephadm 2026-03-08T23:35:58.906182+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Detected new or changed devices on vm02 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: cephadm 2026-03-08T23:35:58.906182+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Detected new or changed devices on vm02 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.912131+0000 mon.a (mon.0) 260 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.912131+0000 mon.a (mon.0) 260 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.916507+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.916507+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.917032+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.917032+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: cephadm 2026-03-08T23:35:58.917374+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Adjusting osd_memory_target on vm02 to 455.7M 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: cephadm 2026-03-08T23:35:58.917374+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Adjusting osd_memory_target on vm02 to 455.7M 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: cephadm 2026-03-08T23:35:58.917729+0000 mgr.a (mgr.14150) 76 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 477921689: error parsing value: Value '477921689' is below minimum 939524096 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: cephadm 2026-03-08T23:35:58.917729+0000 mgr.a (mgr.14150) 76 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 477921689: error parsing value: Value '477921689' is below minimum 939524096 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.917986+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.917986+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.918315+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.918315+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.921394+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:35:59 vm05 bash[22859]: audit 2026-03-08T23:35:58.921394+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: cluster 2026-03-08T23:35:58.570674+0000 mgr.a (mgr.14150) 73 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:00.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: cluster 2026-03-08T23:35:58.570674+0000 mgr.a (mgr.14150) 73 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: cephadm 2026-03-08T23:35:58.906182+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Detected new or changed devices on vm02 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: cephadm 2026-03-08T23:35:58.906182+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Detected new or changed devices on vm02 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.912131+0000 mon.a (mon.0) 260 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.912131+0000 mon.a (mon.0) 260 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.916507+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.916507+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.917032+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.917032+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: cephadm 2026-03-08T23:35:58.917374+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Adjusting osd_memory_target on vm02 to 455.7M 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: cephadm 2026-03-08T23:35:58.917374+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Adjusting osd_memory_target on vm02 to 455.7M 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: cephadm 2026-03-08T23:35:58.917729+0000 mgr.a (mgr.14150) 76 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 477921689: error parsing value: Value '477921689' is below minimum 939524096 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: cephadm 2026-03-08T23:35:58.917729+0000 mgr.a (mgr.14150) 76 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 477921689: error parsing value: Value '477921689' is below minimum 939524096 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.917986+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.917986+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.918315+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.918315+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.921394+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:00.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:35:59 vm02 bash[20612]: audit 2026-03-08T23:35:58.921394+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:02.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:02 vm02 bash[20612]: cluster 2026-03-08T23:36:00.570993+0000 mgr.a (mgr.14150) 77 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:02.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:02 vm02 bash[20612]: cluster 2026-03-08T23:36:00.570993+0000 mgr.a (mgr.14150) 77 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:02.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:02 vm05 bash[22859]: cluster 2026-03-08T23:36:00.570993+0000 mgr.a (mgr.14150) 77 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:02.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:02 vm05 bash[22859]: cluster 2026-03-08T23:36:00.570993+0000 mgr.a (mgr.14150) 77 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:03.565 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.b/config 2026-03-08T23:36:04.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:04 vm02 bash[20612]: cluster 2026-03-08T23:36:02.571285+0000 mgr.a (mgr.14150) 78 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:04.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:04 vm02 bash[20612]: cluster 2026-03-08T23:36:02.571285+0000 mgr.a (mgr.14150) 78 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:04.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:04 vm02 bash[20612]: audit 2026-03-08T23:36:03.811644+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-08T23:36:04.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:04 vm02 bash[20612]: audit 2026-03-08T23:36:03.811644+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-08T23:36:04.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:04 vm02 bash[20612]: audit 2026-03-08T23:36:03.813153+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-08T23:36:04.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:04 vm02 bash[20612]: audit 2026-03-08T23:36:03.813153+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-08T23:36:04.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:04 vm02 bash[20612]: audit 2026-03-08T23:36:03.813680+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:04.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:04 vm02 bash[20612]: audit 2026-03-08T23:36:03.813680+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:04.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:04 vm05 bash[22859]: cluster 2026-03-08T23:36:02.571285+0000 mgr.a (mgr.14150) 78 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:04.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:04 vm05 bash[22859]: cluster 2026-03-08T23:36:02.571285+0000 mgr.a (mgr.14150) 78 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:04.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:04 vm05 bash[22859]: audit 2026-03-08T23:36:03.811644+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-08T23:36:04.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:04 vm05 bash[22859]: audit 2026-03-08T23:36:03.811644+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-08T23:36:04.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:04 vm05 bash[22859]: audit 2026-03-08T23:36:03.813153+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-08T23:36:04.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:04 vm05 bash[22859]: audit 2026-03-08T23:36:03.813153+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-08T23:36:04.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:04 vm05 bash[22859]: audit 2026-03-08T23:36:03.813680+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:04.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:04 vm05 bash[22859]: audit 2026-03-08T23:36:03.813680+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:05.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:05 vm02 bash[20612]: audit 2026-03-08T23:36:03.809870+0000 mgr.a (mgr.14150) 79 : audit [DBG] from='client.24119 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:36:05.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:05 vm02 bash[20612]: audit 2026-03-08T23:36:03.809870+0000 mgr.a (mgr.14150) 79 : audit [DBG] from='client.24119 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:36:05.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:05 vm05 bash[22859]: audit 2026-03-08T23:36:03.809870+0000 mgr.a (mgr.14150) 79 : audit [DBG] from='client.24119 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:36:05.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:05 vm05 bash[22859]: audit 2026-03-08T23:36:03.809870+0000 mgr.a (mgr.14150) 79 : audit [DBG] from='client.24119 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:36:06.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:06 vm05 bash[22859]: cluster 2026-03-08T23:36:04.571499+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:06.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:06 vm05 bash[22859]: cluster 2026-03-08T23:36:04.571499+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:06.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:06 vm02 bash[20612]: cluster 2026-03-08T23:36:04.571499+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:06.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:06 vm02 bash[20612]: cluster 2026-03-08T23:36:04.571499+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:08.587 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:08 vm05 bash[22859]: cluster 2026-03-08T23:36:06.571721+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:08.587 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:08 vm05 bash[22859]: cluster 2026-03-08T23:36:06.571721+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:08.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:08 vm02 bash[20612]: cluster 2026-03-08T23:36:06.571721+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:08.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:08 vm02 bash[20612]: cluster 2026-03-08T23:36:06.571721+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:09.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:08.223501+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.105:0/1589282588' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]: dispatch 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:08.223501+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.105:0/1589282588' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]: dispatch 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:08.242265+0000 mon.a (mon.0) 269 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]: dispatch 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:08.242265+0000 mon.a (mon.0) 269 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]: dispatch 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:08.309506+0000 mon.a (mon.0) 270 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]': finished 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:08.309506+0000 mon.a (mon.0) 270 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]': finished 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: cluster 2026-03-08T23:36:08.311267+0000 mon.a (mon.0) 271 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: cluster 2026-03-08T23:36:08.311267+0000 mon.a (mon.0) 271 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:08.311395+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:08.311395+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:09.040864+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.105:0/3526116414' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-08T23:36:09.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:09 vm05 bash[22859]: audit 2026-03-08T23:36:09.040864+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.105:0/3526116414' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:08.223501+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.105:0/1589282588' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]: dispatch 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:08.223501+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.105:0/1589282588' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]: dispatch 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:08.242265+0000 mon.a (mon.0) 269 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]: dispatch 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:08.242265+0000 mon.a (mon.0) 269 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]: dispatch 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:08.309506+0000 mon.a (mon.0) 270 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]': finished 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:08.309506+0000 mon.a (mon.0) 270 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "19bea9b3-a03a-4de3-be2e-20e893844c4c"}]': finished 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: cluster 2026-03-08T23:36:08.311267+0000 mon.a (mon.0) 271 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: cluster 2026-03-08T23:36:08.311267+0000 mon.a (mon.0) 271 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-08T23:36:09.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:08.311395+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:09.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:08.311395+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:09.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:09.040864+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.105:0/3526116414' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-08T23:36:09.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:09 vm02 bash[20612]: audit 2026-03-08T23:36:09.040864+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.105:0/3526116414' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-08T23:36:10.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:10 vm05 bash[22859]: cluster 2026-03-08T23:36:08.572014+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:10.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:10 vm05 bash[22859]: cluster 2026-03-08T23:36:08.572014+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:10.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:10 vm02 bash[20612]: cluster 2026-03-08T23:36:08.572014+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:10.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:10 vm02 bash[20612]: cluster 2026-03-08T23:36:08.572014+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:12.525 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:12 vm05 bash[22859]: cluster 2026-03-08T23:36:10.572258+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:12.525 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:12 vm05 bash[22859]: cluster 2026-03-08T23:36:10.572258+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:12.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:12 vm02 bash[20612]: cluster 2026-03-08T23:36:10.572258+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:12.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:12 vm02 bash[20612]: cluster 2026-03-08T23:36:10.572258+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:14.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:14 vm05 bash[22859]: cluster 2026-03-08T23:36:12.572552+0000 mgr.a (mgr.14150) 84 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:14.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:14 vm05 bash[22859]: cluster 2026-03-08T23:36:12.572552+0000 mgr.a (mgr.14150) 84 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:14.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:14 vm02 bash[20612]: cluster 2026-03-08T23:36:12.572552+0000 mgr.a (mgr.14150) 84 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:14.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:14 vm02 bash[20612]: cluster 2026-03-08T23:36:12.572552+0000 mgr.a (mgr.14150) 84 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:15.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:15 vm02 bash[20612]: cluster 2026-03-08T23:36:14.572772+0000 mgr.a (mgr.14150) 85 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:15.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:15 vm02 bash[20612]: cluster 2026-03-08T23:36:14.572772+0000 mgr.a (mgr.14150) 85 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:15.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:15 vm05 bash[22859]: cluster 2026-03-08T23:36:14.572772+0000 mgr.a (mgr.14150) 85 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:15.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:15 vm05 bash[22859]: cluster 2026-03-08T23:36:14.572772+0000 mgr.a (mgr.14150) 85 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:17.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:17 vm05 bash[22859]: cluster 2026-03-08T23:36:16.573005+0000 mgr.a (mgr.14150) 86 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:17.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:17 vm05 bash[22859]: cluster 2026-03-08T23:36:16.573005+0000 mgr.a (mgr.14150) 86 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:17.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:17 vm05 bash[22859]: audit 2026-03-08T23:36:17.534539+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-08T23:36:17.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:17 vm05 bash[22859]: audit 2026-03-08T23:36:17.534539+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-08T23:36:17.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:17 vm05 bash[22859]: audit 2026-03-08T23:36:17.535022+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:17.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:17 vm05 bash[22859]: audit 2026-03-08T23:36:17.535022+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:18.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:17 vm02 bash[20612]: cluster 2026-03-08T23:36:16.573005+0000 mgr.a (mgr.14150) 86 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:18.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:17 vm02 bash[20612]: cluster 2026-03-08T23:36:16.573005+0000 mgr.a (mgr.14150) 86 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:18.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:17 vm02 bash[20612]: audit 2026-03-08T23:36:17.534539+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-08T23:36:18.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:17 vm02 bash[20612]: audit 2026-03-08T23:36:17.534539+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-08T23:36:18.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:17 vm02 bash[20612]: audit 2026-03-08T23:36:17.535022+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:18.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:17 vm02 bash[20612]: audit 2026-03-08T23:36:17.535022+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:18.635 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:18 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:36:18.635 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:18 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:36:18.635 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:36:18 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:36:18.635 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:36:18 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:36:18.907 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:18 vm05 bash[22859]: cephadm 2026-03-08T23:36:17.535369+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Deploying daemon osd.1 on vm05 2026-03-08T23:36:18.907 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:18 vm05 bash[22859]: cephadm 2026-03-08T23:36:17.535369+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Deploying daemon osd.1 on vm05 2026-03-08T23:36:19.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:18 vm02 bash[20612]: cephadm 2026-03-08T23:36:17.535369+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Deploying daemon osd.1 on vm05 2026-03-08T23:36:19.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:18 vm02 bash[20612]: cephadm 2026-03-08T23:36:17.535369+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Deploying daemon osd.1 on vm05 2026-03-08T23:36:19.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:19 vm05 bash[22859]: cluster 2026-03-08T23:36:18.573330+0000 mgr.a (mgr.14150) 88 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:19.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:19 vm05 bash[22859]: cluster 2026-03-08T23:36:18.573330+0000 mgr.a (mgr.14150) 88 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:20.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:19 vm05 bash[22859]: audit 2026-03-08T23:36:18.682500+0000 mon.a (mon.0) 275 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:36:20.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:19 vm05 bash[22859]: audit 2026-03-08T23:36:18.682500+0000 mon.a (mon.0) 275 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:36:20.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:19 vm05 bash[22859]: audit 2026-03-08T23:36:18.687559+0000 mon.a (mon.0) 276 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:20.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:19 vm05 bash[22859]: audit 2026-03-08T23:36:18.687559+0000 mon.a (mon.0) 276 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:20.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:19 vm05 bash[22859]: audit 2026-03-08T23:36:18.691963+0000 mon.a (mon.0) 277 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:20.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:19 vm05 bash[22859]: audit 2026-03-08T23:36:18.691963+0000 mon.a (mon.0) 277 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:19 vm02 bash[20612]: cluster 2026-03-08T23:36:18.573330+0000 mgr.a (mgr.14150) 88 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:19 vm02 bash[20612]: cluster 2026-03-08T23:36:18.573330+0000 mgr.a (mgr.14150) 88 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:19 vm02 bash[20612]: audit 2026-03-08T23:36:18.682500+0000 mon.a (mon.0) 275 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:36:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:19 vm02 bash[20612]: audit 2026-03-08T23:36:18.682500+0000 mon.a (mon.0) 275 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:36:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:19 vm02 bash[20612]: audit 2026-03-08T23:36:18.687559+0000 mon.a (mon.0) 276 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:19 vm02 bash[20612]: audit 2026-03-08T23:36:18.687559+0000 mon.a (mon.0) 276 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:19 vm02 bash[20612]: audit 2026-03-08T23:36:18.691963+0000 mon.a (mon.0) 277 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:20.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:19 vm02 bash[20612]: audit 2026-03-08T23:36:18.691963+0000 mon.a (mon.0) 277 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:21.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:21 vm05 bash[22859]: cluster 2026-03-08T23:36:20.573537+0000 mgr.a (mgr.14150) 89 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:21.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:21 vm05 bash[22859]: cluster 2026-03-08T23:36:20.573537+0000 mgr.a (mgr.14150) 89 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:22.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:21 vm02 bash[20612]: cluster 2026-03-08T23:36:20.573537+0000 mgr.a (mgr.14150) 89 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:22.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:21 vm02 bash[20612]: cluster 2026-03-08T23:36:20.573537+0000 mgr.a (mgr.14150) 89 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:22.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:22 vm05 bash[22859]: audit 2026-03-08T23:36:22.525900+0000 mon.a (mon.0) 278 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-08T23:36:22.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:22 vm05 bash[22859]: audit 2026-03-08T23:36:22.525900+0000 mon.a (mon.0) 278 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-08T23:36:23.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:22 vm02 bash[20612]: audit 2026-03-08T23:36:22.525900+0000 mon.a (mon.0) 278 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-08T23:36:23.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:22 vm02 bash[20612]: audit 2026-03-08T23:36:22.525900+0000 mon.a (mon.0) 278 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-08T23:36:23.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: cluster 2026-03-08T23:36:22.573742+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: cluster 2026-03-08T23:36:22.573742+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: audit 2026-03-08T23:36:22.653285+0000 mon.a (mon.0) 279 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: audit 2026-03-08T23:36:22.653285+0000 mon.a (mon.0) 279 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: cluster 2026-03-08T23:36:22.655720+0000 mon.a (mon.0) 280 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: cluster 2026-03-08T23:36:22.655720+0000 mon.a (mon.0) 280 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: audit 2026-03-08T23:36:22.655893+0000 mon.a (mon.0) 281 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: audit 2026-03-08T23:36:22.655893+0000 mon.a (mon.0) 281 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: audit 2026-03-08T23:36:22.655973+0000 mon.a (mon.0) 282 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:24.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:23 vm05 bash[22859]: audit 2026-03-08T23:36:22.655973+0000 mon.a (mon.0) 282 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: cluster 2026-03-08T23:36:22.573742+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: cluster 2026-03-08T23:36:22.573742+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: audit 2026-03-08T23:36:22.653285+0000 mon.a (mon.0) 279 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: audit 2026-03-08T23:36:22.653285+0000 mon.a (mon.0) 279 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: cluster 2026-03-08T23:36:22.655720+0000 mon.a (mon.0) 280 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: cluster 2026-03-08T23:36:22.655720+0000 mon.a (mon.0) 280 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: audit 2026-03-08T23:36:22.655893+0000 mon.a (mon.0) 281 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: audit 2026-03-08T23:36:22.655893+0000 mon.a (mon.0) 281 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-08T23:36:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: audit 2026-03-08T23:36:22.655973+0000 mon.a (mon.0) 282 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:24.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:23 vm02 bash[20612]: audit 2026-03-08T23:36:22.655973+0000 mon.a (mon.0) 282 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:24.908 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:24 vm05 bash[22859]: audit 2026-03-08T23:36:23.656466+0000 mon.a (mon.0) 283 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-08T23:36:24.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:24 vm05 bash[22859]: audit 2026-03-08T23:36:23.656466+0000 mon.a (mon.0) 283 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-08T23:36:24.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:24 vm05 bash[22859]: cluster 2026-03-08T23:36:23.659563+0000 mon.a (mon.0) 284 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-08T23:36:24.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:24 vm05 bash[22859]: cluster 2026-03-08T23:36:23.659563+0000 mon.a (mon.0) 284 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-08T23:36:24.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:24 vm05 bash[22859]: audit 2026-03-08T23:36:23.660696+0000 mon.a (mon.0) 285 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:24.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:24 vm05 bash[22859]: audit 2026-03-08T23:36:23.660696+0000 mon.a (mon.0) 285 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:24.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:24 vm05 bash[22859]: audit 2026-03-08T23:36:23.666698+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:24.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:24 vm05 bash[22859]: audit 2026-03-08T23:36:23.666698+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:24 vm02 bash[20612]: audit 2026-03-08T23:36:23.656466+0000 mon.a (mon.0) 283 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-08T23:36:25.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:24 vm02 bash[20612]: audit 2026-03-08T23:36:23.656466+0000 mon.a (mon.0) 283 : audit [INF] from='osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-08T23:36:25.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:24 vm02 bash[20612]: cluster 2026-03-08T23:36:23.659563+0000 mon.a (mon.0) 284 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-08T23:36:25.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:24 vm02 bash[20612]: cluster 2026-03-08T23:36:23.659563+0000 mon.a (mon.0) 284 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-08T23:36:25.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:24 vm02 bash[20612]: audit 2026-03-08T23:36:23.660696+0000 mon.a (mon.0) 285 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:24 vm02 bash[20612]: audit 2026-03-08T23:36:23.660696+0000 mon.a (mon.0) 285 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:24 vm02 bash[20612]: audit 2026-03-08T23:36:23.666698+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:24 vm02 bash[20612]: audit 2026-03-08T23:36:23.666698+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.899 INFO:teuthology.orchestra.run.vm05.stdout:Created osd(s) 1 on host 'vm05' 2026-03-08T23:36:25.908 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:23.507299+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:23.507299+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:23.507346+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:23.507346+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:24.573956+0000 mgr.a (mgr.14150) 91 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:24.573956+0000 mgr.a (mgr.14150) 91 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:24.663563+0000 mon.a (mon.0) 287 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:24.663563+0000 mon.a (mon.0) 287 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:24.672501+0000 mon.a (mon.0) 288 : cluster [INF] osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868] boot 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:24.672501+0000 mon.a (mon.0) 288 : cluster [INF] osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868] boot 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:24.672512+0000 mon.a (mon.0) 289 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: cluster 2026-03-08T23:36:24.672512+0000 mon.a (mon.0) 289 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:24.672661+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:24.672661+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:24.840722+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:24.840722+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:24.844004+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:24.844004+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:25.253472+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:25.253472+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:25.254190+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:25.254190+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:25.259526+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.909 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:25 vm05 bash[22859]: audit 2026-03-08T23:36:25.259526+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.974 DEBUG:teuthology.orchestra.run.vm05:osd.1> sudo journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@osd.1.service 2026-03-08T23:36:25.975 INFO:tasks.cephadm:Waiting for 2 OSDs to come up... 2026-03-08T23:36:25.975 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph osd stat -f json 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:23.507299+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:23.507299+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:23.507346+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:23.507346+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:24.573956+0000 mgr.a (mgr.14150) 91 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:24.573956+0000 mgr.a (mgr.14150) 91 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:24.663563+0000 mon.a (mon.0) 287 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:24.663563+0000 mon.a (mon.0) 287 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:24.672501+0000 mon.a (mon.0) 288 : cluster [INF] osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868] boot 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:24.672501+0000 mon.a (mon.0) 288 : cluster [INF] osd.1 [v2:192.168.123.105:6800/721622868,v1:192.168.123.105:6801/721622868] boot 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:24.672512+0000 mon.a (mon.0) 289 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: cluster 2026-03-08T23:36:24.672512+0000 mon.a (mon.0) 289 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:24.672661+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:24.672661+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:24.840722+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:24.840722+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:24.844004+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:24.844004+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:25.253472+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:25.253472+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:25.254190+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:25.254190+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:25.259526+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:25.981 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:25 vm02 bash[20612]: audit 2026-03-08T23:36:25.259526+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:26.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:26 vm05 bash[22859]: audit 2026-03-08T23:36:25.888035+0000 mon.a (mon.0) 296 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:36:27.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:26 vm05 bash[22859]: audit 2026-03-08T23:36:25.888035+0000 mon.a (mon.0) 296 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:36:27.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:26 vm05 bash[22859]: audit 2026-03-08T23:36:25.893271+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:27.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:26 vm05 bash[22859]: audit 2026-03-08T23:36:25.893271+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:27.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:26 vm05 bash[22859]: audit 2026-03-08T23:36:25.896614+0000 mon.a (mon.0) 298 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:27.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:26 vm05 bash[22859]: audit 2026-03-08T23:36:25.896614+0000 mon.a (mon.0) 298 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:27.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:26 vm05 bash[22859]: cluster 2026-03-08T23:36:26.274989+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:36:27.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:26 vm05 bash[22859]: cluster 2026-03-08T23:36:26.274989+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:36:27.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:26 vm02 bash[20612]: audit 2026-03-08T23:36:25.888035+0000 mon.a (mon.0) 296 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:36:27.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:26 vm02 bash[20612]: audit 2026-03-08T23:36:25.888035+0000 mon.a (mon.0) 296 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:36:27.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:26 vm02 bash[20612]: audit 2026-03-08T23:36:25.893271+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:27.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:26 vm02 bash[20612]: audit 2026-03-08T23:36:25.893271+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:27.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:26 vm02 bash[20612]: audit 2026-03-08T23:36:25.896614+0000 mon.a (mon.0) 298 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:27.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:26 vm02 bash[20612]: audit 2026-03-08T23:36:25.896614+0000 mon.a (mon.0) 298 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:27.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:26 vm02 bash[20612]: cluster 2026-03-08T23:36:26.274989+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:36:27.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:26 vm02 bash[20612]: cluster 2026-03-08T23:36:26.274989+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:36:28.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:28 vm02 bash[20612]: cluster 2026-03-08T23:36:26.574266+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:28.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:28 vm02 bash[20612]: cluster 2026-03-08T23:36:26.574266+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:28.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:28 vm05 bash[22859]: cluster 2026-03-08T23:36:26.574266+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:28.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:28 vm05 bash[22859]: cluster 2026-03-08T23:36:26.574266+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:30.592 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:30.640 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:30 vm02 bash[20612]: cluster 2026-03-08T23:36:28.574587+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:30.640 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:30 vm02 bash[20612]: cluster 2026-03-08T23:36:28.574587+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:30.697 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:30 vm05 bash[22859]: cluster 2026-03-08T23:36:28.574587+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:30.697 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:30 vm05 bash[22859]: cluster 2026-03-08T23:36:28.574587+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:30.877 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:36:30.929 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"num_osds":2,"num_up_osds":2,"osd_up_since":1773012984,"num_in_osds":2,"osd_in_since":1773012968,"num_remapped_pgs":0} 2026-03-08T23:36:30.929 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph osd dump --format=json 2026-03-08T23:36:31.638 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:31 vm02 bash[20612]: audit 2026-03-08T23:36:30.878294+0000 mon.a (mon.0) 300 : audit [DBG] from='client.? 192.168.123.102:0/1809493736' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-08T23:36:31.638 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:31 vm02 bash[20612]: audit 2026-03-08T23:36:30.878294+0000 mon.a (mon.0) 300 : audit [DBG] from='client.? 192.168.123.102:0/1809493736' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-08T23:36:31.695 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:31 vm05 bash[22859]: audit 2026-03-08T23:36:30.878294+0000 mon.a (mon.0) 300 : audit [DBG] from='client.? 192.168.123.102:0/1809493736' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-08T23:36:31.695 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:31 vm05 bash[22859]: audit 2026-03-08T23:36:30.878294+0000 mon.a (mon.0) 300 : audit [DBG] from='client.? 192.168.123.102:0/1809493736' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: cluster 2026-03-08T23:36:30.574806+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: cluster 2026-03-08T23:36:30.574806+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.505737+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.505737+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.510059+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.510059+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.510909+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.510909+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.511914+0000 mon.a (mon.0) 304 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.511914+0000 mon.a (mon.0) 304 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.512317+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.512317+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.515754+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:32 vm05 bash[22859]: audit 2026-03-08T23:36:31.515754+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: cluster 2026-03-08T23:36:30.574806+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: cluster 2026-03-08T23:36:30.574806+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.505737+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.505737+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.510059+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.510059+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.510909+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.510909+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.511914+0000 mon.a (mon.0) 304 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.511914+0000 mon.a (mon.0) 304 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.512317+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.512317+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.515754+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:32.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:32 vm02 bash[20612]: audit 2026-03-08T23:36:31.515754+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:36:33.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:33 vm05 bash[22859]: cephadm 2026-03-08T23:36:31.500169+0000 mgr.a (mgr.14150) 95 : cephadm [INF] Detected new or changed devices on vm05 2026-03-08T23:36:33.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:33 vm05 bash[22859]: cephadm 2026-03-08T23:36:31.500169+0000 mgr.a (mgr.14150) 95 : cephadm [INF] Detected new or changed devices on vm05 2026-03-08T23:36:33.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:33 vm05 bash[22859]: cephadm 2026-03-08T23:36:31.511285+0000 mgr.a (mgr.14150) 96 : cephadm [INF] Adjusting osd_memory_target on vm05 to 455.7M 2026-03-08T23:36:33.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:33 vm05 bash[22859]: cephadm 2026-03-08T23:36:31.511285+0000 mgr.a (mgr.14150) 96 : cephadm [INF] Adjusting osd_memory_target on vm05 to 455.7M 2026-03-08T23:36:33.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:33 vm05 bash[22859]: cephadm 2026-03-08T23:36:31.511631+0000 mgr.a (mgr.14150) 97 : cephadm [WRN] Unable to set osd_memory_target on vm05 to 477915955: error parsing value: Value '477915955' is below minimum 939524096 2026-03-08T23:36:33.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:33 vm05 bash[22859]: cephadm 2026-03-08T23:36:31.511631+0000 mgr.a (mgr.14150) 97 : cephadm [WRN] Unable to set osd_memory_target on vm05 to 477915955: error parsing value: Value '477915955' is below minimum 939524096 2026-03-08T23:36:33.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:33 vm02 bash[20612]: cephadm 2026-03-08T23:36:31.500169+0000 mgr.a (mgr.14150) 95 : cephadm [INF] Detected new or changed devices on vm05 2026-03-08T23:36:33.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:33 vm02 bash[20612]: cephadm 2026-03-08T23:36:31.500169+0000 mgr.a (mgr.14150) 95 : cephadm [INF] Detected new or changed devices on vm05 2026-03-08T23:36:33.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:33 vm02 bash[20612]: cephadm 2026-03-08T23:36:31.511285+0000 mgr.a (mgr.14150) 96 : cephadm [INF] Adjusting osd_memory_target on vm05 to 455.7M 2026-03-08T23:36:33.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:33 vm02 bash[20612]: cephadm 2026-03-08T23:36:31.511285+0000 mgr.a (mgr.14150) 96 : cephadm [INF] Adjusting osd_memory_target on vm05 to 455.7M 2026-03-08T23:36:33.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:33 vm02 bash[20612]: cephadm 2026-03-08T23:36:31.511631+0000 mgr.a (mgr.14150) 97 : cephadm [WRN] Unable to set osd_memory_target on vm05 to 477915955: error parsing value: Value '477915955' is below minimum 939524096 2026-03-08T23:36:33.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:33 vm02 bash[20612]: cephadm 2026-03-08T23:36:31.511631+0000 mgr.a (mgr.14150) 97 : cephadm [WRN] Unable to set osd_memory_target on vm05 to 477915955: error parsing value: Value '477915955' is below minimum 939524096 2026-03-08T23:36:34.603 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:34.616 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:34 vm02 bash[20612]: cluster 2026-03-08T23:36:32.575033+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:34.616 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:34 vm02 bash[20612]: cluster 2026-03-08T23:36:32.575033+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:34.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:34 vm05 bash[22859]: cluster 2026-03-08T23:36:32.575033+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:34.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:34 vm05 bash[22859]: cluster 2026-03-08T23:36:32.575033+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:34.840 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:36:34.840 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"fsid":"32c51f16-1b47-11f1-bdef-c3bc34602109","created":"2026-03-08T23:34:14.089697+0000","modified":"2026-03-08T23:36:26.257936+0000","last_up_change":"2026-03-08T23:36:24.662659+0000","last_in_change":"2026-03-08T23:36:08.242542+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"6663b6e9-7873-428d-8f01-0747ce138477","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6803","nonce":1666513430}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6805","nonce":1666513430}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6809","nonce":1666513430}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6807","nonce":1666513430}]},"public_addr":"192.168.123.102:6803/1666513430","cluster_addr":"192.168.123.102:6805/1666513430","heartbeat_back_addr":"192.168.123.102:6809/1666513430","heartbeat_front_addr":"192.168.123.102:6807/1666513430","state":["exists","up"]},{"osd":1,"uuid":"19bea9b3-a03a-4de3-be2e-20e893844c4c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6801","nonce":721622868}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6803","nonce":721622868}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6807","nonce":721622868}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6805","nonce":721622868}]},"public_addr":"192.168.123.105:6801/721622868","cluster_addr":"192.168.123.105:6803/721622868","heartbeat_back_addr":"192.168.123.105:6807/721622868","heartbeat_front_addr":"192.168.123.105:6805/721622868","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-08T23:35:50.129595+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-08T23:36:23.507348+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/1684559568":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6800/139367019":"2026-03-09T23:34:34.552017+0000","192.168.123.102:0/418052460":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6801/139367019":"2026-03-09T23:34:34.552017+0000","192.168.123.102:0/2609492251":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6801/2416289666":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/1338323692":"2026-03-09T23:34:25.014439+0000","192.168.123.102:6800/2416289666":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/2522337660":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/2030697492":"2026-03-09T23:34:25.014439+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-08T23:36:34.888 INFO:tasks.cephadm.ceph_manager.ceph:[] 2026-03-08T23:36:34.888 INFO:tasks.cephadm:Setting up client nodes... 2026-03-08T23:36:34.888 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-08T23:36:34.888 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-08T23:36:34.888 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph mgr dump --format=json 2026-03-08T23:36:35.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:35 vm05 bash[22859]: audit 2026-03-08T23:36:34.841279+0000 mon.a (mon.0) 307 : audit [DBG] from='client.? 192.168.123.102:0/1485123383' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:35.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:35 vm05 bash[22859]: audit 2026-03-08T23:36:34.841279+0000 mon.a (mon.0) 307 : audit [DBG] from='client.? 192.168.123.102:0/1485123383' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:35.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:35 vm02 bash[20612]: audit 2026-03-08T23:36:34.841279+0000 mon.a (mon.0) 307 : audit [DBG] from='client.? 192.168.123.102:0/1485123383' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:35.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:35 vm02 bash[20612]: audit 2026-03-08T23:36:34.841279+0000 mon.a (mon.0) 307 : audit [DBG] from='client.? 192.168.123.102:0/1485123383' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:36.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:36 vm05 bash[22859]: cluster 2026-03-08T23:36:34.575308+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:36.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:36 vm05 bash[22859]: cluster 2026-03-08T23:36:34.575308+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:36.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:36 vm02 bash[20612]: cluster 2026-03-08T23:36:34.575308+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:36.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:36 vm02 bash[20612]: cluster 2026-03-08T23:36:34.575308+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:38.615 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:38.632 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:38 vm02 bash[20612]: cluster 2026-03-08T23:36:36.575564+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:38.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:38 vm02 bash[20612]: cluster 2026-03-08T23:36:36.575564+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:38.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:38 vm05 bash[22859]: cluster 2026-03-08T23:36:36.575564+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:38.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:38 vm05 bash[22859]: cluster 2026-03-08T23:36:36.575564+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:38.888 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:36:38.936 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":13,"flags":0,"active_gid":14150,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":1809341858},{"type":"v1","addr":"192.168.123.102:6801","nonce":1809341858}]},"active_addr":"192.168.123.102:6801/1809341858","active_change":"2026-03-08T23:34:34.552265+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":24105,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.102:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":3187554829}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":165651757}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":1625248140}]}]} 2026-03-08T23:36:38.938 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-08T23:36:38.938 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-08T23:36:38.938 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph osd dump --format=json 2026-03-08T23:36:39.630 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:39 vm02 bash[20612]: audit 2026-03-08T23:36:38.886851+0000 mon.a (mon.0) 308 : audit [DBG] from='client.? 192.168.123.102:0/3877789730' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-08T23:36:39.631 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:39 vm02 bash[20612]: audit 2026-03-08T23:36:38.886851+0000 mon.a (mon.0) 308 : audit [DBG] from='client.? 192.168.123.102:0/3877789730' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-08T23:36:39.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:39 vm05 bash[22859]: audit 2026-03-08T23:36:38.886851+0000 mon.a (mon.0) 308 : audit [DBG] from='client.? 192.168.123.102:0/3877789730' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-08T23:36:39.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:39 vm05 bash[22859]: audit 2026-03-08T23:36:38.886851+0000 mon.a (mon.0) 308 : audit [DBG] from='client.? 192.168.123.102:0/3877789730' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-08T23:36:40.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:40 vm05 bash[22859]: cluster 2026-03-08T23:36:38.575834+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:40.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:40 vm05 bash[22859]: cluster 2026-03-08T23:36:38.575834+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:40.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:40 vm02 bash[20612]: cluster 2026-03-08T23:36:38.575834+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:40.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:40 vm02 bash[20612]: cluster 2026-03-08T23:36:38.575834+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:42.627 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:42.642 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:42 vm02 bash[20612]: cluster 2026-03-08T23:36:40.576071+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:42.642 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:42 vm02 bash[20612]: cluster 2026-03-08T23:36:40.576071+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:42.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:42 vm05 bash[22859]: cluster 2026-03-08T23:36:40.576071+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:42.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:42 vm05 bash[22859]: cluster 2026-03-08T23:36:40.576071+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:42.866 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:36:42.866 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"fsid":"32c51f16-1b47-11f1-bdef-c3bc34602109","created":"2026-03-08T23:34:14.089697+0000","modified":"2026-03-08T23:36:26.257936+0000","last_up_change":"2026-03-08T23:36:24.662659+0000","last_in_change":"2026-03-08T23:36:08.242542+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"6663b6e9-7873-428d-8f01-0747ce138477","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6803","nonce":1666513430}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6805","nonce":1666513430}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6809","nonce":1666513430}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6807","nonce":1666513430}]},"public_addr":"192.168.123.102:6803/1666513430","cluster_addr":"192.168.123.102:6805/1666513430","heartbeat_back_addr":"192.168.123.102:6809/1666513430","heartbeat_front_addr":"192.168.123.102:6807/1666513430","state":["exists","up"]},{"osd":1,"uuid":"19bea9b3-a03a-4de3-be2e-20e893844c4c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6801","nonce":721622868}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6803","nonce":721622868}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6807","nonce":721622868}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6805","nonce":721622868}]},"public_addr":"192.168.123.105:6801/721622868","cluster_addr":"192.168.123.105:6803/721622868","heartbeat_back_addr":"192.168.123.105:6807/721622868","heartbeat_front_addr":"192.168.123.105:6805/721622868","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-08T23:35:50.129595+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-08T23:36:23.507348+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/1684559568":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6800/139367019":"2026-03-09T23:34:34.552017+0000","192.168.123.102:0/418052460":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6801/139367019":"2026-03-09T23:34:34.552017+0000","192.168.123.102:0/2609492251":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6801/2416289666":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/1338323692":"2026-03-09T23:34:25.014439+0000","192.168.123.102:6800/2416289666":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/2522337660":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/2030697492":"2026-03-09T23:34:25.014439+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-08T23:36:42.919 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-08T23:36:42.919 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph osd dump --format=json 2026-03-08T23:36:43.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:43 vm05 bash[22859]: audit 2026-03-08T23:36:42.867405+0000 mon.a (mon.0) 309 : audit [DBG] from='client.? 192.168.123.102:0/877395410' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:43.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:43 vm05 bash[22859]: audit 2026-03-08T23:36:42.867405+0000 mon.a (mon.0) 309 : audit [DBG] from='client.? 192.168.123.102:0/877395410' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:43.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:43 vm02 bash[20612]: audit 2026-03-08T23:36:42.867405+0000 mon.a (mon.0) 309 : audit [DBG] from='client.? 192.168.123.102:0/877395410' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:43.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:43 vm02 bash[20612]: audit 2026-03-08T23:36:42.867405+0000 mon.a (mon.0) 309 : audit [DBG] from='client.? 192.168.123.102:0/877395410' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:44.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:44 vm05 bash[22859]: cluster 2026-03-08T23:36:42.576371+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:44.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:44 vm05 bash[22859]: cluster 2026-03-08T23:36:42.576371+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:44.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:44 vm02 bash[20612]: cluster 2026-03-08T23:36:42.576371+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:44.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:44 vm02 bash[20612]: cluster 2026-03-08T23:36:42.576371+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:46.637 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:46.654 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:46 vm02 bash[20612]: cluster 2026-03-08T23:36:44.576687+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:46.654 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:46 vm02 bash[20612]: cluster 2026-03-08T23:36:44.576687+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:46.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:46 vm05 bash[22859]: cluster 2026-03-08T23:36:44.576687+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:46.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:46 vm05 bash[22859]: cluster 2026-03-08T23:36:44.576687+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:46.922 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:36:46.922 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"fsid":"32c51f16-1b47-11f1-bdef-c3bc34602109","created":"2026-03-08T23:34:14.089697+0000","modified":"2026-03-08T23:36:26.257936+0000","last_up_change":"2026-03-08T23:36:24.662659+0000","last_in_change":"2026-03-08T23:36:08.242542+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"6663b6e9-7873-428d-8f01-0747ce138477","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6803","nonce":1666513430}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6805","nonce":1666513430}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6809","nonce":1666513430}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1666513430},{"type":"v1","addr":"192.168.123.102:6807","nonce":1666513430}]},"public_addr":"192.168.123.102:6803/1666513430","cluster_addr":"192.168.123.102:6805/1666513430","heartbeat_back_addr":"192.168.123.102:6809/1666513430","heartbeat_front_addr":"192.168.123.102:6807/1666513430","state":["exists","up"]},{"osd":1,"uuid":"19bea9b3-a03a-4de3-be2e-20e893844c4c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6801","nonce":721622868}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6803","nonce":721622868}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6807","nonce":721622868}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":721622868},{"type":"v1","addr":"192.168.123.105:6805","nonce":721622868}]},"public_addr":"192.168.123.105:6801/721622868","cluster_addr":"192.168.123.105:6803/721622868","heartbeat_back_addr":"192.168.123.105:6807/721622868","heartbeat_front_addr":"192.168.123.105:6805/721622868","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-08T23:35:50.129595+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-08T23:36:23.507348+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/1684559568":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6800/139367019":"2026-03-09T23:34:34.552017+0000","192.168.123.102:0/418052460":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6801/139367019":"2026-03-09T23:34:34.552017+0000","192.168.123.102:0/2609492251":"2026-03-09T23:34:34.552017+0000","192.168.123.102:6801/2416289666":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/1338323692":"2026-03-09T23:34:25.014439+0000","192.168.123.102:6800/2416289666":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/2522337660":"2026-03-09T23:34:25.014439+0000","192.168.123.102:0/2030697492":"2026-03-09T23:34:25.014439+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-08T23:36:46.976 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph tell osd.0 flush_pg_stats 2026-03-08T23:36:46.976 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph tell osd.1 flush_pg_stats 2026-03-08T23:36:47.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:47 vm05 bash[22859]: audit 2026-03-08T23:36:46.923123+0000 mon.a (mon.0) 310 : audit [DBG] from='client.? 192.168.123.102:0/2653334405' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:47.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:47 vm05 bash[22859]: audit 2026-03-08T23:36:46.923123+0000 mon.a (mon.0) 310 : audit [DBG] from='client.? 192.168.123.102:0/2653334405' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:47.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:47 vm02 bash[20612]: audit 2026-03-08T23:36:46.923123+0000 mon.a (mon.0) 310 : audit [DBG] from='client.? 192.168.123.102:0/2653334405' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:47.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:47 vm02 bash[20612]: audit 2026-03-08T23:36:46.923123+0000 mon.a (mon.0) 310 : audit [DBG] from='client.? 192.168.123.102:0/2653334405' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-08T23:36:48.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:48 vm05 bash[22859]: cluster 2026-03-08T23:36:46.576922+0000 mgr.a (mgr.14150) 105 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:48.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:48 vm05 bash[22859]: cluster 2026-03-08T23:36:46.576922+0000 mgr.a (mgr.14150) 105 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:48.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:48 vm02 bash[20612]: cluster 2026-03-08T23:36:46.576922+0000 mgr.a (mgr.14150) 105 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:48.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:48 vm02 bash[20612]: cluster 2026-03-08T23:36:46.576922+0000 mgr.a (mgr.14150) 105 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:50.648 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:50.648 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:50.701 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:50 vm02 bash[20612]: cluster 2026-03-08T23:36:48.577178+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:50.701 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:50 vm02 bash[20612]: cluster 2026-03-08T23:36:48.577178+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:50.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:50 vm05 bash[22859]: cluster 2026-03-08T23:36:48.577178+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:50.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:50 vm05 bash[22859]: cluster 2026-03-08T23:36:48.577178+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:50.872 INFO:teuthology.orchestra.run.vm02.stdout:55834574855 2026-03-08T23:36:50.872 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph osd last-stat-seq osd.1 2026-03-08T23:36:50.925 INFO:teuthology.orchestra.run.vm02.stdout:34359738381 2026-03-08T23:36:50.925 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph osd last-stat-seq osd.0 2026-03-08T23:36:51.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:51 vm05 bash[22859]: cluster 2026-03-08T23:36:50.577382+0000 mgr.a (mgr.14150) 107 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:51.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:51 vm05 bash[22859]: cluster 2026-03-08T23:36:50.577382+0000 mgr.a (mgr.14150) 107 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:51.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:51 vm02 bash[20612]: cluster 2026-03-08T23:36:50.577382+0000 mgr.a (mgr.14150) 107 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:51.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:51 vm02 bash[20612]: cluster 2026-03-08T23:36:50.577382+0000 mgr.a (mgr.14150) 107 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:53.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:53 vm05 bash[22859]: cluster 2026-03-08T23:36:52.577625+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:53.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:53 vm05 bash[22859]: cluster 2026-03-08T23:36:52.577625+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:54.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:53 vm02 bash[20612]: cluster 2026-03-08T23:36:52.577625+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:54.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:53 vm02 bash[20612]: cluster 2026-03-08T23:36:52.577625+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:54.660 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:54.660 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:54.921 INFO:teuthology.orchestra.run.vm02.stdout:34359738382 2026-03-08T23:36:54.925 INFO:teuthology.orchestra.run.vm02.stdout:55834574855 2026-03-08T23:36:54.984 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738381 got 34359738382 for osd.0 2026-03-08T23:36:54.984 DEBUG:teuthology.parallel:result is None 2026-03-08T23:36:55.011 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574855 got 55834574855 for osd.1 2026-03-08T23:36:55.011 DEBUG:teuthology.parallel:result is None 2026-03-08T23:36:55.011 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-08T23:36:55.011 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph pg dump --format=json 2026-03-08T23:36:55.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:55 vm05 bash[22859]: cluster 2026-03-08T23:36:54.577828+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:55.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:55 vm05 bash[22859]: cluster 2026-03-08T23:36:54.577828+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:55.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:55 vm05 bash[22859]: audit 2026-03-08T23:36:54.921666+0000 mon.a (mon.0) 311 : audit [DBG] from='client.? 192.168.123.102:0/3447021255' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-08T23:36:55.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:55 vm05 bash[22859]: audit 2026-03-08T23:36:54.921666+0000 mon.a (mon.0) 311 : audit [DBG] from='client.? 192.168.123.102:0/3447021255' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-08T23:36:55.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:55 vm05 bash[22859]: audit 2026-03-08T23:36:54.926459+0000 mon.a (mon.0) 312 : audit [DBG] from='client.? 192.168.123.102:0/3224841262' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-08T23:36:55.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:55 vm05 bash[22859]: audit 2026-03-08T23:36:54.926459+0000 mon.a (mon.0) 312 : audit [DBG] from='client.? 192.168.123.102:0/3224841262' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-08T23:36:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:55 vm02 bash[20612]: cluster 2026-03-08T23:36:54.577828+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:55 vm02 bash[20612]: cluster 2026-03-08T23:36:54.577828+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:55 vm02 bash[20612]: audit 2026-03-08T23:36:54.921666+0000 mon.a (mon.0) 311 : audit [DBG] from='client.? 192.168.123.102:0/3447021255' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-08T23:36:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:55 vm02 bash[20612]: audit 2026-03-08T23:36:54.921666+0000 mon.a (mon.0) 311 : audit [DBG] from='client.? 192.168.123.102:0/3447021255' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-08T23:36:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:55 vm02 bash[20612]: audit 2026-03-08T23:36:54.926459+0000 mon.a (mon.0) 312 : audit [DBG] from='client.? 192.168.123.102:0/3224841262' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-08T23:36:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:55 vm02 bash[20612]: audit 2026-03-08T23:36:54.926459+0000 mon.a (mon.0) 312 : audit [DBG] from='client.? 192.168.123.102:0/3224841262' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-08T23:36:57.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:57 vm05 bash[22859]: cluster 2026-03-08T23:36:56.578032+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:58.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:57 vm05 bash[22859]: cluster 2026-03-08T23:36:56.578032+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:58.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:57 vm02 bash[20612]: cluster 2026-03-08T23:36:56.578032+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:58.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:57 vm02 bash[20612]: cluster 2026-03-08T23:36:56.578032+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:59.670 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:36:59.909 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:36:59.910 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-08T23:36:59.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:59 vm02 bash[20612]: cluster 2026-03-08T23:36:58.578295+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:59.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:36:59 vm02 bash[20612]: cluster 2026-03-08T23:36:58.578295+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:59.959 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":76,"stamp":"2026-03-08T23:36:58.578190+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":53920,"kb_used_data":240,"kb_used_omap":3,"kb_used_meta":53628,"kb_avail":41880928,"statfs":{"total":42941284352,"available":42886070272,"internally_reserved":0,"allocated":245760,"data_stored":60148,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":3184,"internal_metadata":54915984},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":13,"seq":55834574856,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738383,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1594,"internal_metadata":27457990},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-08T23:36:59.959 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph pg dump --format=json 2026-03-08T23:36:59.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:59 vm05 bash[22859]: cluster 2026-03-08T23:36:58.578295+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:36:59.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:36:59 vm05 bash[22859]: cluster 2026-03-08T23:36:58.578295+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:00.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:00 vm05 bash[22859]: audit 2026-03-08T23:36:59.911077+0000 mgr.a (mgr.14150) 112 : audit [DBG] from='client.14262 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:00.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:00 vm05 bash[22859]: audit 2026-03-08T23:36:59.911077+0000 mgr.a (mgr.14150) 112 : audit [DBG] from='client.14262 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:01.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:00 vm02 bash[20612]: audit 2026-03-08T23:36:59.911077+0000 mgr.a (mgr.14150) 112 : audit [DBG] from='client.14262 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:01.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:00 vm02 bash[20612]: audit 2026-03-08T23:36:59.911077+0000 mgr.a (mgr.14150) 112 : audit [DBG] from='client.14262 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:01.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:01 vm05 bash[22859]: cluster 2026-03-08T23:37:00.578550+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:01.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:01 vm05 bash[22859]: cluster 2026-03-08T23:37:00.578550+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:02.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:01 vm02 bash[20612]: cluster 2026-03-08T23:37:00.578550+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:02.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:01 vm02 bash[20612]: cluster 2026-03-08T23:37:00.578550+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:03.681 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:03.938 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:37:03.938 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-08T23:37:03.948 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:03 vm02 bash[20612]: cluster 2026-03-08T23:37:02.578903+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:03.948 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:03 vm02 bash[20612]: cluster 2026-03-08T23:37:02.578903+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:03.988 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":78,"stamp":"2026-03-08T23:37:02.578773+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":53920,"kb_used_data":240,"kb_used_omap":3,"kb_used_meta":53628,"kb_avail":41880928,"statfs":{"total":42941284352,"available":42886070272,"internally_reserved":0,"allocated":245760,"data_stored":60148,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":3184,"internal_metadata":54915984},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":13,"seq":55834574857,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738383,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1594,"internal_metadata":27457990},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-08T23:37:03.988 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-08T23:37:03.988 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-08T23:37:03.988 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-08T23:37:03.988 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph health --format=json 2026-03-08T23:37:03.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:03 vm05 bash[22859]: cluster 2026-03-08T23:37:02.578903+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:03.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:03 vm05 bash[22859]: cluster 2026-03-08T23:37:02.578903+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:04.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:04 vm05 bash[22859]: audit 2026-03-08T23:37:03.939592+0000 mgr.a (mgr.14150) 115 : audit [DBG] from='client.14266 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:04.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:04 vm05 bash[22859]: audit 2026-03-08T23:37:03.939592+0000 mgr.a (mgr.14150) 115 : audit [DBG] from='client.14266 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:05.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:04 vm02 bash[20612]: audit 2026-03-08T23:37:03.939592+0000 mgr.a (mgr.14150) 115 : audit [DBG] from='client.14266 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:05.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:04 vm02 bash[20612]: audit 2026-03-08T23:37:03.939592+0000 mgr.a (mgr.14150) 115 : audit [DBG] from='client.14266 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:05.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:05 vm05 bash[22859]: cluster 2026-03-08T23:37:04.579153+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:05.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:05 vm05 bash[22859]: cluster 2026-03-08T23:37:04.579153+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:06.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:05 vm02 bash[20612]: cluster 2026-03-08T23:37:04.579153+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:06.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:05 vm02 bash[20612]: cluster 2026-03-08T23:37:04.579153+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:07.690 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:07.949 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:37:07.950 INFO:teuthology.orchestra.run.vm02.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-08T23:37:07.959 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:07 vm02 bash[20612]: cluster 2026-03-08T23:37:06.579367+0000 mgr.a (mgr.14150) 117 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:07.959 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:07 vm02 bash[20612]: cluster 2026-03-08T23:37:06.579367+0000 mgr.a (mgr.14150) 117 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:07.998 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-08T23:37:07.998 INFO:tasks.cephadm:Setup complete, yielding 2026-03-08T23:37:07.998 INFO:teuthology.run_tasks:Running task exec... 2026-03-08T23:37:08.001 INFO:teuthology.task.exec:Executing custom commands... 2026-03-08T23:37:08.001 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm02.local 2026-03-08T23:37:08.001 DEBUG:teuthology.orchestra.run.vm02:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-08T23:37:08.001 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:07 vm05 bash[22859]: cluster 2026-03-08T23:37:06.579367+0000 mgr.a (mgr.14150) 117 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:08.002 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:07 vm05 bash[22859]: cluster 2026-03-08T23:37:06.579367+0000 mgr.a (mgr.14150) 117 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:08.009 INFO:teuthology.task.exec:Running commands on role host.b host ubuntu@vm05.local 2026-03-08T23:37:08.009 DEBUG:teuthology.orchestra.run.vm05:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-08T23:37:08.017 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-08T23:37:08.020 INFO:tasks.cephadm:Applying spec(s): extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon --- custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest --- custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z 2026-03-08T23:37:08.020 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch apply -i - 2026-03-08T23:37:08.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:08 vm05 bash[22859]: audit 2026-03-08T23:37:07.951233+0000 mon.a (mon.0) 313 : audit [DBG] from='client.? 192.168.123.102:0/725444051' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-08T23:37:08.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:08 vm05 bash[22859]: audit 2026-03-08T23:37:07.951233+0000 mon.a (mon.0) 313 : audit [DBG] from='client.? 192.168.123.102:0/725444051' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-08T23:37:09.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:08 vm02 bash[20612]: audit 2026-03-08T23:37:07.951233+0000 mon.a (mon.0) 313 : audit [DBG] from='client.? 192.168.123.102:0/725444051' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-08T23:37:09.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:08 vm02 bash[20612]: audit 2026-03-08T23:37:07.951233+0000 mon.a (mon.0) 313 : audit [DBG] from='client.? 192.168.123.102:0/725444051' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-08T23:37:09.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:09 vm05 bash[22859]: cluster 2026-03-08T23:37:08.579617+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:09.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:09 vm05 bash[22859]: cluster 2026-03-08T23:37:08.579617+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:10.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:09 vm02 bash[20612]: cluster 2026-03-08T23:37:08.579617+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:10.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:09 vm02 bash[20612]: cluster 2026-03-08T23:37:08.579617+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:11.701 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:11.979 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled mon update... 2026-03-08T23:37:11.979 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled container.foo update... 2026-03-08T23:37:11.979 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled container.bar update... 2026-03-08T23:37:11.988 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:11 vm02 bash[20612]: cluster 2026-03-08T23:37:10.579821+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:11.988 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:11 vm02 bash[20612]: cluster 2026-03-08T23:37:10.579821+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:11.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:11 vm05 bash[22859]: cluster 2026-03-08T23:37:10.579821+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:11.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:11 vm05 bash[22859]: cluster 2026-03-08T23:37:10.579821+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:12.050 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-08T23:37:12.052 INFO:tasks.cephadm:Waiting for ceph service mon to start (timeout 300)... 2026-03-08T23:37:12.052 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch ls -f json 2026-03-08T23:37:12.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:12.798 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:37:12 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:12.798 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:37:12 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.945856+0000 mgr.a (mgr.14150) 120 : audit [DBG] from='client.14274 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.945856+0000 mgr.a (mgr.14150) 120 : audit [DBG] from='client.14274 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: cephadm 2026-03-08T23:37:11.952779+0000 mgr.a (mgr.14150) 121 : cephadm [INF] Saving service mon spec with placement * 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: cephadm 2026-03-08T23:37:11.952779+0000 mgr.a (mgr.14150) 121 : cephadm [INF] Saving service mon spec with placement * 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.956986+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.956986+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: cephadm 2026-03-08T23:37:11.957742+0000 mgr.a (mgr.14150) 122 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: cephadm 2026-03-08T23:37:11.957742+0000 mgr.a (mgr.14150) 122 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.958003+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.958003+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.961823+0000 mon.a (mon.0) 316 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.961823+0000 mon.a (mon.0) 316 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: cephadm 2026-03-08T23:37:11.964951+0000 mgr.a (mgr.14150) 123 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: cephadm 2026-03-08T23:37:11.964951+0000 mgr.a (mgr.14150) 123 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.968942+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:11.968942+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:12.286828+0000 mon.a (mon.0) 318 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:12.286828+0000 mon.a (mon.0) 318 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:12.287423+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:12.287423+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:12.291931+0000 mon.a (mon.0) 320 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:12.291931+0000 mon.a (mon.0) 320 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:12.295229+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: audit 2026-03-08T23:37:12.295229+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: cephadm 2026-03-08T23:37:12.296347+0000 mgr.a (mgr.14150) 124 : cephadm [INF] Deploying daemon container.foo.vm02 on vm02 2026-03-08T23:37:13.230 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:12 vm05 bash[22859]: cephadm 2026-03-08T23:37:12.296347+0000 mgr.a (mgr.14150) 124 : cephadm [INF] Deploying daemon container.foo.vm02 on vm02 2026-03-08T23:37:13.297 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:37:12 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.297 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:37:12 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.945856+0000 mgr.a (mgr.14150) 120 : audit [DBG] from='client.14274 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.945856+0000 mgr.a (mgr.14150) 120 : audit [DBG] from='client.14274 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: cephadm 2026-03-08T23:37:11.952779+0000 mgr.a (mgr.14150) 121 : cephadm [INF] Saving service mon spec with placement * 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: cephadm 2026-03-08T23:37:11.952779+0000 mgr.a (mgr.14150) 121 : cephadm [INF] Saving service mon spec with placement * 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.956986+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.956986+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: cephadm 2026-03-08T23:37:11.957742+0000 mgr.a (mgr.14150) 122 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: cephadm 2026-03-08T23:37:11.957742+0000 mgr.a (mgr.14150) 122 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.958003+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.958003+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.961823+0000 mon.a (mon.0) 316 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.961823+0000 mon.a (mon.0) 316 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: cephadm 2026-03-08T23:37:11.964951+0000 mgr.a (mgr.14150) 123 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: cephadm 2026-03-08T23:37:11.964951+0000 mgr.a (mgr.14150) 123 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.968942+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:11.968942+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:12.286828+0000 mon.a (mon.0) 318 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:12.286828+0000 mon.a (mon.0) 318 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:12.287423+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:12.287423+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:12.291931+0000 mon.a (mon.0) 320 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:12.291931+0000 mon.a (mon.0) 320 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:12.295229+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: audit 2026-03-08T23:37:12.295229+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: cephadm 2026-03-08T23:37:12.296347+0000 mgr.a (mgr.14150) 124 : cephadm [INF] Deploying daemon container.foo.vm02 on vm02 2026-03-08T23:37:13.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:12 vm02 bash[20612]: cephadm 2026-03-08T23:37:12.296347+0000 mgr.a (mgr.14150) 124 : cephadm [INF] Deploying daemon container.foo.vm02 on vm02 2026-03-08T23:37:13.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.500 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:37:13 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.500 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:37:13 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.940 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.941 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:37:13 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:13.941 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:37:13 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: cluster 2026-03-08T23:37:12.580051+0000 mgr.a (mgr.14150) 125 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: cluster 2026-03-08T23:37:12.580051+0000 mgr.a (mgr.14150) 125 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:12.975589+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:12.975589+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:12.979756+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:12.979756+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:12.983801+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:12.983801+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: cephadm 2026-03-08T23:37:12.984608+0000 mgr.a (mgr.14150) 126 : cephadm [INF] Deploying daemon container.foo.vm05 on vm05 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: cephadm 2026-03-08T23:37:12.984608+0000 mgr.a (mgr.14150) 126 : cephadm [INF] Deploying daemon container.foo.vm05 on vm05 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:13.680411+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:13.680411+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:13.684708+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:13.684708+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:13.689095+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:13.689095+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:13.692708+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:13 vm05 bash[22859]: audit 2026-03-08T23:37:13.692708+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:14 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:14.250 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:37:14 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:14.250 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:37:14 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: cluster 2026-03-08T23:37:12.580051+0000 mgr.a (mgr.14150) 125 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: cluster 2026-03-08T23:37:12.580051+0000 mgr.a (mgr.14150) 125 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:12.975589+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:12.975589+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:12.979756+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:12.979756+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:12.983801+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:12.983801+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: cephadm 2026-03-08T23:37:12.984608+0000 mgr.a (mgr.14150) 126 : cephadm [INF] Deploying daemon container.foo.vm05 on vm05 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: cephadm 2026-03-08T23:37:12.984608+0000 mgr.a (mgr.14150) 126 : cephadm [INF] Deploying daemon container.foo.vm05 on vm05 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:13.680411+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:13.680411+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:13.684708+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:13.684708+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:13.689095+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:13.689095+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:13.692708+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:13 vm02 bash[20612]: audit 2026-03-08T23:37:13.692708+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:14.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:14 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:14.750 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:37:14 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:14.750 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:37:14 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:15.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:14 vm02 bash[20612]: cephadm 2026-03-08T23:37:13.693679+0000 mgr.a (mgr.14150) 127 : cephadm [INF] Deploying daemon container.bar.vm05 on vm05 2026-03-08T23:37:15.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:14 vm02 bash[20612]: cephadm 2026-03-08T23:37:13.693679+0000 mgr.a (mgr.14150) 127 : cephadm [INF] Deploying daemon container.bar.vm05 on vm05 2026-03-08T23:37:15.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:14 vm02 bash[20612]: audit 2026-03-08T23:37:14.610546+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:15.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:14 vm02 bash[20612]: audit 2026-03-08T23:37:14.610546+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:15.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:14 vm05 bash[22859]: cephadm 2026-03-08T23:37:13.693679+0000 mgr.a (mgr.14150) 127 : cephadm [INF] Deploying daemon container.bar.vm05 on vm05 2026-03-08T23:37:15.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:14 vm05 bash[22859]: cephadm 2026-03-08T23:37:13.693679+0000 mgr.a (mgr.14150) 127 : cephadm [INF] Deploying daemon container.bar.vm05 on vm05 2026-03-08T23:37:15.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:15 vm05 bash[22859]: audit 2026-03-08T23:37:14.610546+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:15.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:15 vm05 bash[22859]: audit 2026-03-08T23:37:14.610546+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:15.716 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:16.084 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:37:16.084 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-08T23:37:11.975089Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-08T23:37:11.964955Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-08T23:37:13.692908Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-08T23:37:11.957843Z", "running": 0, "size": 2}}, {"events": ["2026-03-08T23:35:19.611525Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-08T23:35:18.822236Z", "last_refresh": "2026-03-08T23:35:51.983578Z", "running": 2, "size": 2}}, {"events": ["2026-03-08T23:37:12.295461Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-08T23:37:11.952794Z", "last_refresh": "2026-03-08T23:35:51.983660Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-08T23:35:51.983628Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-08T23:37:16.171 INFO:tasks.cephadm:mon has 2/2 2026-03-08T23:37:16.171 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-08T23:37:16.174 INFO:tasks.cephadm:Waiting for ceph service container.foo to start (timeout 300)... 2026-03-08T23:37:16.174 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch ls -f json 2026-03-08T23:37:16.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:15 vm05 bash[22859]: cluster 2026-03-08T23:37:14.580316+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:16.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:15 vm05 bash[22859]: cluster 2026-03-08T23:37:14.580316+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:16.301 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:15 vm02 bash[20612]: cluster 2026-03-08T23:37:14.580316+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:16.301 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:15 vm02 bash[20612]: cluster 2026-03-08T23:37:14.580316+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:17.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:17 vm02 bash[20612]: audit 2026-03-08T23:37:16.084696+0000 mgr.a (mgr.14150) 129 : audit [DBG] from='client.14278 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:17.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:17 vm02 bash[20612]: audit 2026-03-08T23:37:16.084696+0000 mgr.a (mgr.14150) 129 : audit [DBG] from='client.14278 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:17.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:17 vm05 bash[22859]: audit 2026-03-08T23:37:16.084696+0000 mgr.a (mgr.14150) 129 : audit [DBG] from='client.14278 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:17.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:17 vm05 bash[22859]: audit 2026-03-08T23:37:16.084696+0000 mgr.a (mgr.14150) 129 : audit [DBG] from='client.14278 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:18.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:18 vm05 bash[22859]: cluster 2026-03-08T23:37:16.580543+0000 mgr.a (mgr.14150) 130 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:18.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:18 vm05 bash[22859]: cluster 2026-03-08T23:37:16.580543+0000 mgr.a (mgr.14150) 130 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:18.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:18 vm02 bash[20612]: cluster 2026-03-08T23:37:16.580543+0000 mgr.a (mgr.14150) 130 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:18.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:18 vm02 bash[20612]: cluster 2026-03-08T23:37:16.580543+0000 mgr.a (mgr.14150) 130 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:20.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:20 vm02 bash[20612]: cluster 2026-03-08T23:37:18.580811+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:20.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:20 vm02 bash[20612]: cluster 2026-03-08T23:37:18.580811+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:20.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:20 vm05 bash[22859]: cluster 2026-03-08T23:37:18.580811+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:20.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:20 vm05 bash[22859]: cluster 2026-03-08T23:37:18.580811+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:21.813 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:22.220 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:22.220 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:37:22 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:22.221 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:37:22 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:22.240 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:37:22.240 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-08T23:37:21.599319Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-08T23:37:11.964955Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-08T23:37:13.692908Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-08T23:37:11.957843Z", "running": 0, "size": 2}}, {"events": ["2026-03-08T23:35:19.611525Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-08T23:35:18.822236Z", "last_refresh": "2026-03-08T23:35:51.983578Z", "running": 2, "size": 2}}, {"events": ["2026-03-08T23:37:12.295461Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-08T23:37:11.952794Z", "last_refresh": "2026-03-08T23:35:51.983660Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-08T23:35:51.983628Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-08T23:37:22.339 INFO:tasks.cephadm:container.foo has 0/2 2026-03-08T23:37:22.547 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:37:22 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:22.547 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:37:22 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 bash[20612]: cluster 2026-03-08T23:37:20.580995+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 bash[20612]: cluster 2026-03-08T23:37:20.580995+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 bash[20612]: audit 2026-03-08T23:37:21.590528+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 bash[20612]: audit 2026-03-08T23:37:21.590528+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 bash[20612]: audit 2026-03-08T23:37:21.595010+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 bash[20612]: audit 2026-03-08T23:37:21.595010+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 bash[20612]: audit 2026-03-08T23:37:21.598992+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:22 vm02 bash[20612]: audit 2026-03-08T23:37:21.598992+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:22 vm05 bash[22859]: cluster 2026-03-08T23:37:20.580995+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:22.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:22 vm05 bash[22859]: cluster 2026-03-08T23:37:20.580995+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:22.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:22 vm05 bash[22859]: audit 2026-03-08T23:37:21.590528+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:22 vm05 bash[22859]: audit 2026-03-08T23:37:21.590528+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:22 vm05 bash[22859]: audit 2026-03-08T23:37:21.595010+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:22 vm05 bash[22859]: audit 2026-03-08T23:37:21.595010+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:22 vm05 bash[22859]: audit 2026-03-08T23:37:21.598992+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:22.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:22 vm05 bash[22859]: audit 2026-03-08T23:37:21.598992+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:23.339 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch ls -f json 2026-03-08T23:37:23.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:23 vm05 bash[22859]: cephadm 2026-03-08T23:37:21.599830+0000 mgr.a (mgr.14150) 133 : cephadm [INF] Deploying daemon container.bar.vm02 on vm02 2026-03-08T23:37:23.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:23 vm05 bash[22859]: cephadm 2026-03-08T23:37:21.599830+0000 mgr.a (mgr.14150) 133 : cephadm [INF] Deploying daemon container.bar.vm02 on vm02 2026-03-08T23:37:23.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:23 vm05 bash[22859]: audit 2026-03-08T23:37:22.240651+0000 mgr.a (mgr.14150) 134 : audit [DBG] from='client.14282 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:23.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:23 vm05 bash[22859]: audit 2026-03-08T23:37:22.240651+0000 mgr.a (mgr.14150) 134 : audit [DBG] from='client.14282 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:23.667 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:23 vm02 bash[20612]: cephadm 2026-03-08T23:37:21.599830+0000 mgr.a (mgr.14150) 133 : cephadm [INF] Deploying daemon container.bar.vm02 on vm02 2026-03-08T23:37:23.667 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:23 vm02 bash[20612]: cephadm 2026-03-08T23:37:21.599830+0000 mgr.a (mgr.14150) 133 : cephadm [INF] Deploying daemon container.bar.vm02 on vm02 2026-03-08T23:37:23.667 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:23 vm02 bash[20612]: audit 2026-03-08T23:37:22.240651+0000 mgr.a (mgr.14150) 134 : audit [DBG] from='client.14282 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:23.667 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:23 vm02 bash[20612]: audit 2026-03-08T23:37:22.240651+0000 mgr.a (mgr.14150) 134 : audit [DBG] from='client.14282 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:24.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:24 vm05 bash[22859]: cluster 2026-03-08T23:37:22.581243+0000 mgr.a (mgr.14150) 135 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:24.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:24 vm05 bash[22859]: cluster 2026-03-08T23:37:22.581243+0000 mgr.a (mgr.14150) 135 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:24.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:24 vm02 bash[20612]: cluster 2026-03-08T23:37:22.581243+0000 mgr.a (mgr.14150) 135 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:24.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:24 vm02 bash[20612]: cluster 2026-03-08T23:37:22.581243+0000 mgr.a (mgr.14150) 135 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:26.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:26 vm05 bash[22859]: cluster 2026-03-08T23:37:24.581455+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:26.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:26 vm05 bash[22859]: cluster 2026-03-08T23:37:24.581455+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:26.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:26 vm02 bash[20612]: cluster 2026-03-08T23:37:24.581455+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:26.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:26 vm02 bash[20612]: cluster 2026-03-08T23:37:24.581455+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:27.958 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:28.979 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:28 vm02 bash[20612]: cluster 2026-03-08T23:37:26.581704+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:28.979 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:28 vm02 bash[20612]: cluster 2026-03-08T23:37:26.581704+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:28.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:28 vm05 bash[22859]: cluster 2026-03-08T23:37:26.581704+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:28.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:28 vm05 bash[22859]: cluster 2026-03-08T23:37:26.581704+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:29.233 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:37:29.233 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-08T23:37:21.599319Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-08T23:37:11.964955Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-08T23:37:13.692908Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-08T23:37:11.957843Z", "running": 0, "size": 2}}, {"events": ["2026-03-08T23:35:19.611525Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-08T23:35:18.822236Z", "last_refresh": "2026-03-08T23:35:51.983578Z", "running": 2, "size": 2}}, {"events": ["2026-03-08T23:37:12.295461Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-08T23:37:11.952794Z", "last_refresh": "2026-03-08T23:35:51.983660Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-08T23:35:51.983628Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-08T23:37:29.331 INFO:tasks.cephadm:container.foo has 0/2 2026-03-08T23:37:29.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: cluster 2026-03-08T23:37:28.581943+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: cluster 2026-03-08T23:37:28.581943+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.231535+0000 mgr.a (mgr.14150) 139 : audit [DBG] from='client.14286 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.231535+0000 mgr.a (mgr.14150) 139 : audit [DBG] from='client.14286 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.539114+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.539114+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.546168+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.546168+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.554603+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.554603+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.560966+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.560966+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.579571+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.579571+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.633614+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:29 vm02 bash[20612]: audit 2026-03-08T23:37:29.633614+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:29.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: cluster 2026-03-08T23:37:28.581943+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:29.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: cluster 2026-03-08T23:37:28.581943+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:29.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.231535+0000 mgr.a (mgr.14150) 139 : audit [DBG] from='client.14286 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.231535+0000 mgr.a (mgr.14150) 139 : audit [DBG] from='client.14286 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.539114+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.539114+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.546168+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.546168+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.554603+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.554603+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.560966+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.560966+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.579571+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.579571+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.633614+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:29 vm05 bash[22859]: audit 2026-03-08T23:37:29.633614+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:30.332 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch ls -f json 2026-03-08T23:37:31.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:31 vm02 bash[20612]: cluster 2026-03-08T23:37:30.582132+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:31.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:31 vm02 bash[20612]: cluster 2026-03-08T23:37:30.582132+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:31.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:31 vm05 bash[22859]: cluster 2026-03-08T23:37:30.582132+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:31.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:31 vm05 bash[22859]: cluster 2026-03-08T23:37:30.582132+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:33.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:33 vm05 bash[22859]: cluster 2026-03-08T23:37:32.582397+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:33.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:33 vm05 bash[22859]: cluster 2026-03-08T23:37:32.582397+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:34.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:33 vm02 bash[20612]: cluster 2026-03-08T23:37:32.582397+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:34.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:33 vm02 bash[20612]: cluster 2026-03-08T23:37:32.582397+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:34.369 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:34.677 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:37:34.677 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-08T23:37:29.561653Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-08T23:37:11.964955Z", "last_refresh": "2026-03-08T23:37:34.541943Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-08T23:37:13.692908Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-08T23:37:11.957843Z", "last_refresh": "2026-03-08T23:37:34.541879Z", "running": 2, "size": 2}}, {"events": ["2026-03-08T23:35:19.611525Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-08T23:35:18.822236Z", "last_refresh": "2026-03-08T23:37:34.542036Z", "running": 2, "size": 2}}, {"events": ["2026-03-08T23:37:12.295461Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-08T23:37:11.952794Z", "last_refresh": "2026-03-08T23:37:34.541978Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-08T23:37:34.542007Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-08T23:37:34.726 INFO:tasks.cephadm:container.foo has 2/2 2026-03-08T23:37:34.726 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-08T23:37:34.728 INFO:tasks.cephadm:Waiting for ceph service container.bar to start (timeout 300)... 2026-03-08T23:37:34.728 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph orch ls -f json 2026-03-08T23:37:35.423 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:35.423 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:35.424 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:35.680 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:35.680 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: Stopping Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:37:35.681 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[20612]: debug 2026-03-08T23:37:35.464+0000 7f319676d640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-08T23:37:35.681 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[20612]: debug 2026-03-08T23:37:35.464+0000 7f319676d640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-08T23:37:35.681 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36879]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mon-a 2026-03-08T23:37:35.681 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service: Deactivated successfully. 2026-03-08T23:37:35.681 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: Stopped Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:37:36.047 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 systemd[1]: Started Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.868+0000 7f83a6f66d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.868+0000 7f83a6f66d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.868+0000 7f83a6f66d80 0 pidfile_write: ignore empty --pid-file 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.868+0000 7f83a6f66d80 0 load: jerasure load: lrc 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Git sha 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: DB SUMMARY 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: DB Session ID: TD5DKPA2JHF29LETWGMG 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: CURRENT file: CURRENT 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: MANIFEST file: MANIFEST-000015 size: 281 Bytes 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000014.log size: 3739694 ; 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.error_if_exists: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.create_if_missing: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.env: 0x55c3e6738dc0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.info_log: 0x55c3fb119880 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.statistics: (nil) 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.use_fsync: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.db_log_dir: 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.wal_dir: 2026-03-08T23:37:36.049 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.write_buffer_manager: 0x55c3fb11d900 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.unordered_write: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.row_cache: None 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.wal_filter: None 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.two_write_queues: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.wal_compression: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.atomic_flush: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_open_files: -1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Compression algorithms supported: 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: kZSTD supported: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: kXpressCompression supported: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: kZlibCompression supported: 1 2026-03-08T23:37:36.050 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.merge_operator: 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_filter: None 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55c3fb118480) 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: cache_index_and_filter_blocks: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: pin_top_level_index_and_filter: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: index_type: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: data_block_index_type: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: index_shortening: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: data_block_hash_table_util_ratio: 0.750000 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: checksum: 4 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: no_block_cache: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: block_cache: 0x55c3fb13f350 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: block_cache_name: BinnedLRUCache 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: block_cache_options: 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: capacity : 536870912 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: num_shard_bits : 4 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: strict_capacity_limit : 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: high_pri_pool_ratio: 0.000 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: block_cache_compressed: (nil) 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: persistent_cache: (nil) 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: block_size: 4096 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: block_size_deviation: 10 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: block_restart_interval: 16 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: index_block_restart_interval: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: metadata_block_size: 4096 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: partition_filters: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: use_delta_encoding: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: filter_policy: bloomfilter 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: whole_key_filtering: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: verify_compression: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: read_amp_bytes_per_bit: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: format_version: 5 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: enable_index_compression: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: block_align: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: max_auto_readahead_size: 262144 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: prepopulate_block_cache: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: initial_auto_readahead_size: 8192 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: num_file_reads_for_auto_readahead: 2 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression: NoCompression 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.num_levels: 7 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-08T23:37:36.051 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.bloom_locality: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-08T23:37:36.052 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.ttl: 2592000 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.enable_blob_files: false 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.min_blob_size: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.872+0000 7f83a6f66d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.876+0000 7f83a6f66d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 225, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.876+0000 7f83a6f66d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.876+0000 7f83a6f66d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: bf9a7919-2cb0-4dde-8761-0dace08ae027 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.876+0000 7f83a6f66d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013055880067, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.876+0000 7f83a6f66d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.900+0000 7f83a6f66d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013055904302, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 3280220, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 230, "largest_seqno": 3269, "table_properties": {"data_size": 3269658, "index_size": 6193, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3333, "raw_key_size": 29462, "raw_average_key_size": 22, "raw_value_size": 3244503, "raw_average_value_size": 2439, "num_data_blocks": 294, "num_entries": 1330, "num_filter_entries": 1330, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773013055, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "bf9a7919-2cb0-4dde-8761-0dace08ae027", "db_session_id": "TD5DKPA2JHF29LETWGMG", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.900+0000 7f83a6f66d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013055904448, "job": 1, "event": "recovery_finished"} 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.900+0000 7f83a6f66d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.912+0000 7f83a6f66d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.912+0000 7f83a6f66d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55c3fb140e00 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.912+0000 7f83a6f66d80 4 rocksdb: DB pointer 0x55c3fb24e000 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.916+0000 7f839cd30640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: debug 2026-03-08T23:37:35.916+0000 7f839cd30640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: ** DB Stats ** 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:36.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: ** Compaction Stats [default] ** 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: L0 3/0 3.20 MB 0.8 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 182.3 0.02 0.00 1 0.017 0 0 0.0 0.0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Sum 3/0 3.20 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 182.3 0.02 0.00 1 0.017 0 0 0.0 0.0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 182.3 0.02 0.00 1 0.017 0 0 0.0 0.0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: ** Compaction Stats [default] ** 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 182.3 0.02 0.00 1 0.017 0 0 0.0 0.0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Flush(GB): cumulative 0.003, interval 0.003 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: AddFile(Total Files): cumulative 0, interval 0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: AddFile(Keys): cumulative 0, interval 0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Cumulative compaction: 0.00 GB write, 77.42 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Interval compaction: 0.00 GB write, 77.42 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Block cache BinnedLRUCache@0x55c3fb13f350#7 capacity: 512.00 MB usage: 46.94 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.7e-05 secs_since: 0 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: Block cache entry stats(count,size,portion): DataBlock(5,35.27 KB,0.00672638%) FilterBlock(3,4.73 KB,0.00090301%) IndexBlock(3,6.94 KB,0.00132322%) Misc(1,0.00 KB,0%) 2026-03-08T23:37:36.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:35 vm02 bash[36992]: ** File Read Latency Histogram By Level [default] ** 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:35.473586+0000 mon.b (mon.1) 8 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:35.473586+0000 mon.b (mon.1) 8 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:35.473750+0000 mon.b (mon.1) 9 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:35.473750+0000 mon.b (mon.1) 9 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.124325+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.124325+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.132579+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.132579+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140551+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140551+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140558+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140558+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140561+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140561+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140564+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140564+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140566+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140566+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140568+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140568+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140571+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140571+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140573+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140573+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140884+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140884+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140895+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.140895+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.145975+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.145975+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.146110+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: cluster 2026-03-08T23:37:36.146110+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.164564+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.164564+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.172608+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.172608+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.176538+0000 mon.b (mon.1) 10 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.176538+0000 mon.b (mon.1) 10 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.177088+0000 mon.b (mon.1) 11 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.177088+0000 mon.b (mon.1) 11 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.177606+0000 mon.b (mon.1) 12 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:36.461 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: audit 2026-03-08T23:37:36.177606+0000 mon.b (mon.1) 12 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:35.473586+0000 mon.b (mon.1) 8 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:35.473586+0000 mon.b (mon.1) 8 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:35.473750+0000 mon.b (mon.1) 9 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:35.473750+0000 mon.b (mon.1) 9 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.124325+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.124325+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.132579+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.132579+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140551+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140551+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140558+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140558+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140561+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140561+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140564+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140564+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140566+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140566+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140568+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140568+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140571+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140571+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140573+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140573+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140884+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140884+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140895+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.140895+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.145975+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.145975+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.146110+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: cluster 2026-03-08T23:37:36.146110+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.164564+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.164564+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.172608+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.172608+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.176538+0000 mon.b (mon.1) 10 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.176538+0000 mon.b (mon.1) 10 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.177088+0000 mon.b (mon.1) 11 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.177088+0000 mon.b (mon.1) 11 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.177606+0000 mon.b (mon.1) 12 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:36.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:36 vm02 bash[36992]: audit 2026-03-08T23:37:36.177606+0000 mon.b (mon.1) 12 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:36.717 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:36.717 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 systemd[1]: Stopping Ceph mon.b for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:37:36.717 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:37:36 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:36.718 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:37:36 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:36.979 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: debug 2026-03-08T23:37:36.745+0000 7fe277ae1640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-08T23:37:36.979 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[22859]: debug 2026-03-08T23:37:36.745+0000 7fe277ae1640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-08T23:37:37.249 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:37:37 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:37.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:36 vm05 bash[30398]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mon-b 2026-03-08T23:37:37.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 systemd[1]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.b.service: Deactivated successfully. 2026-03-08T23:37:37.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 systemd[1]: Stopped Ceph mon.b for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:37:37.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:37.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 systemd[1]: Started Ceph mon.b for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:37:37.250 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:37:37 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 0 pidfile_write: ignore empty --pid-file 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 0 load: jerasure load: lrc 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Git sha 0 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: DB SUMMARY 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: DB Session ID: WCUKSIFEKC5VJR0KEZ61 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: CURRENT file: CURRENT 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 1, files: 000008.sst 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000009.log size: 4678582 ; 2026-03-08T23:37:37.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.error_if_exists: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.create_if_missing: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.env: 0x55895da46dc0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.info_log: 0x55895e7ead00 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.statistics: (nil) 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.301+0000 7f8efe79fd80 4 rocksdb: Options.use_fsync: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.db_log_dir: 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.wal_dir: 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.write_buffer_manager: 0x55895e7ef900 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.unordered_write: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.row_cache: None 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.wal_filter: None 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.two_write_queues: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.wal_compression: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.atomic_flush: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-08T23:37:37.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_open_files: -1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Compression algorithms supported: 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: kZSTD supported: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: kXpressCompression supported: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: kZlibCompression supported: 1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.merge_operator: 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_filter: None 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55895e7ea480) 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: cache_index_and_filter_blocks: 1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: pin_top_level_index_and_filter: 1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: index_type: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: data_block_index_type: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: index_shortening: 1 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: data_block_hash_table_util_ratio: 0.750000 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: checksum: 4 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: no_block_cache: 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: block_cache: 0x55895e811350 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: block_cache_name: BinnedLRUCache 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: block_cache_options: 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: capacity : 536870912 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: num_shard_bits : 4 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: strict_capacity_limit : 0 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: high_pri_pool_ratio: 0.000 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: block_cache_compressed: (nil) 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: persistent_cache: (nil) 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: block_size: 4096 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: block_size_deviation: 10 2026-03-08T23:37:37.752 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: block_restart_interval: 16 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: index_block_restart_interval: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: metadata_block_size: 4096 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: partition_filters: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: use_delta_encoding: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: filter_policy: bloomfilter 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: whole_key_filtering: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: verify_compression: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: read_amp_bytes_per_bit: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: format_version: 5 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: enable_index_compression: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: block_align: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: max_auto_readahead_size: 262144 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: prepopulate_block_cache: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: initial_auto_readahead_size: 8192 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: num_file_reads_for_auto_readahead: 2 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression: NoCompression 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.num_levels: 7 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-08T23:37:37.753 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.bloom_locality: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.ttl: 2592000 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.enable_blob_files: false 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.min_blob_size: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.305+0000 7f8efe79fd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.309+0000 7f8efe79fd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.309+0000 7f8efe79fd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.309+0000 7f8efe79fd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 7ce5a3f0-68c5-45fd-bad2-ce95fe1c5ad6 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.309+0000 7f8efe79fd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013057312183, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.309+0000 7f8efe79fd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.325+0000 7f8efe79fd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013057326310, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 3388891, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 6, "largest_seqno": 3242, "table_properties": {"data_size": 3377527, "index_size": 6483, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3845, "raw_key_size": 39319, "raw_average_key_size": 25, "raw_value_size": 3346986, "raw_average_value_size": 2183, "num_data_blocks": 306, "num_entries": 1533, "num_filter_entries": 1533, "num_deletions": 4, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773013057, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "7ce5a3f0-68c5-45fd-bad2-ce95fe1c5ad6", "db_session_id": "WCUKSIFEKC5VJR0KEZ61", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.325+0000 7f8efe79fd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013057326436, "job": 1, "event": "recovery_finished"} 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.325+0000 7f8efe79fd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.325+0000 7f8efe79fd80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.325+0000 7f8efe79fd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55895e812e00 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.325+0000 7f8efe79fd80 4 rocksdb: DB pointer 0x55895e920000 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8ef4569640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8ef4569640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-08T23:37:37.754 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: ** DB Stats ** 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: ** Compaction Stats [default] ** 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: L0 2/0 3.23 MB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 336.1 0.01 0.00 1 0.010 0 0 0.0 0.0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Sum 2/0 3.23 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 336.1 0.01 0.00 1 0.010 0 0 0.0 0.0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 336.1 0.01 0.00 1 0.010 0 0 0.0 0.0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: ** Compaction Stats [default] ** 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 336.1 0.01 0.00 1 0.010 0 0 0.0 0.0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Flush(GB): cumulative 0.003, interval 0.003 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: AddFile(Total Files): cumulative 0, interval 0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: AddFile(Keys): cumulative 0, interval 0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Cumulative compaction: 0.00 GB write, 136.14 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Interval compaction: 0.00 GB write, 136.14 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Block cache BinnedLRUCache@0x55895e811350#7 capacity: 512.00 MB usage: 11.47 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.7e-05 secs_since: 0 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: Block cache entry stats(count,size,portion): FilterBlock(2,4.14 KB,0.000789762%) IndexBlock(2,6.69 KB,0.00127554%) Misc(2,0.64 KB,0.00012219%) 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: ** File Read Latency Histogram By Level [default] ** 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 0 starting mon.b rank 1 at public addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] at bind addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 1 mon.b@-1(???) e2 preinit fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 0 mon.b@-1(???).mds e1 new map 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 0 mon.b@-1(???).mds e1 print_map 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: e1 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: btime 2026-03-08T23:34:14:089156+0000 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: legacy client fscid: -1 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: No filesystems configured 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 0 mon.b@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:37.755 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:37 vm05 bash[30512]: debug 2026-03-08T23:37:37.329+0000 7f8efe79fd80 1 mon.b@-1(???).paxosservice(auth 1..8) refresh upgraded, format 0 -> 3 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.338651+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.338651+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.343038+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.343038+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.345712+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.345712+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349845+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349845+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349864+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349864+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349875+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349875+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349886+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349886+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349895+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349895+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349904+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349904+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349912+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349912+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349920+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.349920+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.350241+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.350241+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.350261+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.350261+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.350470+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.350470+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.350547+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: cluster 2026-03-08T23:37:37.350547+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: audit 2026-03-08T23:37:37.353488+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: audit 2026-03-08T23:37:37.353488+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: audit 2026-03-08T23:37:37.356640+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:38.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: audit 2026-03-08T23:37:37.356640+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:38.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: audit 2026-03-08T23:37:37.357477+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:38.751 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:38 vm05 bash[30512]: audit 2026-03-08T23:37:37.357477+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.338651+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.338651+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.343038+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.343038+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.345712+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.345712+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349845+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349845+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349864+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349864+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349875+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349875+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349886+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349886+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349895+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349895+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349904+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349904+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349912+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349912+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349920+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:38.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.349920+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.350241+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.350241+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.350261+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.350261+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.350470+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.350470+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.350547+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: cluster 2026-03-08T23:37:37.350547+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: audit 2026-03-08T23:37:37.353488+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: audit 2026-03-08T23:37:37.353488+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: audit 2026-03-08T23:37:37.356640+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: audit 2026-03-08T23:37:37.356640+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: audit 2026-03-08T23:37:37.357477+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:38.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:38 vm02 bash[36992]: audit 2026-03-08T23:37:37.357477+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:39.402 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:39.636 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-08T23:37:39.636 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-08T23:37:29.561653Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-08T23:37:11.964955Z", "last_refresh": "2026-03-08T23:37:34.541943Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-08T23:37:13.692908Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-08T23:37:11.957843Z", "last_refresh": "2026-03-08T23:37:34.541879Z", "running": 2, "size": 2}}, {"events": ["2026-03-08T23:35:19.611525Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm05=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-08T23:35:18.822236Z", "last_refresh": "2026-03-08T23:37:34.542036Z", "running": 2, "size": 2}}, {"events": ["2026-03-08T23:37:12.295461Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-08T23:37:11.952794Z", "running": 0, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", "container_image_name": "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc", "last_refresh": "2026-03-08T23:37:34.542007Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-08T23:37:39.683 INFO:tasks.cephadm:container.bar has 2/2 2026-03-08T23:37:39.683 INFO:teuthology.run_tasks:Running task exec... 2026-03-08T23:37:39.685 INFO:teuthology.task.exec:Executing custom commands... 2026-03-08T23:37:39.685 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm02.local 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'set -ex 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> sleep 60 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # check extra container and entrypoint args written to mon unit run file 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # check that custom container properly wrote content to file. 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # This requires the custom config, extra container args, and 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # entrypoint args to all be working in order for this to have 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # been written. The container entrypoint was set up with custom_configs, 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # the content and where to write to with the entrypoint args, and the mounting 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # of the /etc/cephadm_testing dir with extra container args 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> grep "testing_custom_containers" /etc/cephadm_testing/testing.txt 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # Verify that container bar'"'"'s init containers and primary container 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> # ran successfully 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name '"'"'container.bar.*'"'"') 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> test -n "$dir" 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> grep ok ${dir}/data/primary.txt 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> grep from=centos ${dir}/data/from.txt 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> test -s ${dir}/data/presized.dat 2026-03-08T23:37:39.685 DEBUG:teuthology.orchestra.run.vm02:> ' 2026-03-08T23:37:39.692 INFO:teuthology.orchestra.run.vm02.stderr:++ /home/ubuntu/cephtest/cephadm shell -- ceph fsid 2026-03-08T23:37:40.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:40 vm05 bash[30512]: cluster 2026-03-08T23:37:38.583534+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:40.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:40 vm05 bash[30512]: cluster 2026-03-08T23:37:38.583534+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:40.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:40 vm02 bash[36992]: cluster 2026-03-08T23:37:38.583534+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:40.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:40 vm02 bash[36992]: cluster 2026-03-08T23:37:38.583534+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:41.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:41 vm05 bash[30512]: audit 2026-03-08T23:37:39.636282+0000 mgr.a (mgr.14150) 150 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:41.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:41 vm05 bash[30512]: audit 2026-03-08T23:37:39.636282+0000 mgr.a (mgr.14150) 150 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:41.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:41 vm02 bash[36992]: audit 2026-03-08T23:37:39.636282+0000 mgr.a (mgr.14150) 150 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:41.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:41 vm02 bash[36992]: audit 2026-03-08T23:37:39.636282+0000 mgr.a (mgr.14150) 150 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: cluster 2026-03-08T23:37:40.583750+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: cluster 2026-03-08T23:37:40.583750+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:41.590211+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:41.590211+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:41.594430+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:41.594430+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.280277+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.280277+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.285111+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.285111+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.286058+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.286058+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.286475+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.286475+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.289925+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.289925+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.300235+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.300235+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.300646+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.300646+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.301023+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:42.633 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: audit 2026-03-08T23:37:42.301023+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: cluster 2026-03-08T23:37:40.583750+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: cluster 2026-03-08T23:37:40.583750+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:41.590211+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:41.590211+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:41.594430+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:41.594430+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.280277+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.280277+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.285111+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.285111+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.286058+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.286058+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.286475+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.286475+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.289925+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.289925+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.300235+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.300235+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.300646+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.300646+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.301023+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:42.750 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:42 vm05 bash[30512]: audit 2026-03-08T23:37:42.301023+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:43.047 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:37:42 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:43.047 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:37:42 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:43.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:43.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 systemd[1]: Stopping Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:37:43.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: debug 2026-03-08T23:37:42.872+0000 7f83a5541640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-08T23:37:43.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:42 vm02 bash[36992]: debug 2026-03-08T23:37:42.872+0000 7f83a5541640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-08T23:37:43.388 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:37:43 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:43.388 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 08 23:37:43 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:43.388 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37463]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mon-a 2026-03-08T23:37:43.389 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 systemd[1]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service: Deactivated successfully. 2026-03-08T23:37:43.389 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 systemd[1]: Stopped Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:37:43.389 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:43.389 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 systemd[1]: Started Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:37:43.425 INFO:teuthology.orchestra.run.vm02.stderr:Inferring fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:43.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.420+0000 7f516f25fd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-08T23:37:43.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.420+0000 7f516f25fd80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-08T23:37:43.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.420+0000 7f516f25fd80 0 pidfile_write: ignore empty --pid-file 2026-03-08T23:37:43.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 0 load: jerasure load: lrc 2026-03-08T23:37:43.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-08T23:37:43.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Git sha 0 2026-03-08T23:37:43.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: DB SUMMARY 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: DB Session ID: VFFS8P9RT4UFSVMWNQK0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: CURRENT file: CURRENT 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: MANIFEST file: MANIFEST-000020 size: 373 Bytes 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 3, files: 000008.sst 000013.sst 000018.sst 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000019.log size: 223991 ; 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.error_if_exists: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.create_if_missing: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.env: 0x56532a0addc0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.info_log: 0x5653337a9880 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.statistics: (nil) 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.use_fsync: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.db_log_dir: 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.wal_dir: 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.write_buffer_manager: 0x5653337ad900 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.unordered_write: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.row_cache: None 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.wal_filter: None 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.two_write_queues: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.424+0000 7f516f25fd80 4 rocksdb: Options.wal_compression: 0 2026-03-08T23:37:43.800 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.atomic_flush: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_open_files: -1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Compression algorithms supported: 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: kZSTD supported: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: kXpressCompression supported: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: kZlibCompression supported: 1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.merge_operator: 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_filter: None 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5653337a94e0) 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: cache_index_and_filter_blocks: 1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: pin_top_level_index_and_filter: 1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: index_type: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: data_block_index_type: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: index_shortening: 1 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: data_block_hash_table_util_ratio: 0.750000 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: checksum: 4 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: no_block_cache: 0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: block_cache: 0x5653337ce9b0 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: block_cache_name: BinnedLRUCache 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: block_cache_options: 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: capacity : 536870912 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: num_shard_bits : 4 2026-03-08T23:37:43.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: strict_capacity_limit : 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: high_pri_pool_ratio: 0.000 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: block_cache_compressed: (nil) 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: persistent_cache: (nil) 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: block_size: 4096 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: block_size_deviation: 10 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: block_restart_interval: 16 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: index_block_restart_interval: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: metadata_block_size: 4096 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: partition_filters: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: use_delta_encoding: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: filter_policy: bloomfilter 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: whole_key_filtering: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: verify_compression: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: read_amp_bytes_per_bit: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: format_version: 5 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: enable_index_compression: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: block_align: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: max_auto_readahead_size: 262144 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: prepopulate_block_cache: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: initial_auto_readahead_size: 8192 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: num_file_reads_for_auto_readahead: 2 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression: NoCompression 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.num_levels: 7 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-08T23:37:43.802 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.bloom_locality: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.ttl: 2592000 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.enable_blob_files: false 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.min_blob_size: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 succeeded,manifest_file_number is 20, next_file_number is 22, last_sequence is 3269, log_number is 15,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 15 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 15 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: bf9a7919-2cb0-4dde-8761-0dace08ae027 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013063434599, "job": 1, "event": "recovery_started", "wal_files": [19]} 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.428+0000 7f516f25fd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #19 mode 2 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.432+0000 7f516f25fd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013063437028, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 23, "file_size": 200004, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 3275, "largest_seqno": 3470, "table_properties": {"data_size": 197930, "index_size": 721, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 325, "raw_key_size": 2697, "raw_average_key_size": 24, "raw_value_size": 195449, "raw_average_value_size": 1793, "num_data_blocks": 31, "num_entries": 109, "num_filter_entries": 109, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773013063, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "bf9a7919-2cb0-4dde-8761-0dace08ae027", "db_session_id": "VFFS8P9RT4UFSVMWNQK0", "orig_file_number": 23, "seqno_to_time_mapping": "N/A"}} 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.432+0000 7f516f25fd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013063437087, "job": 1, "event": "recovery_finished"} 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.432+0000 7f516f25fd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 25 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f516f25fd80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000019.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f516f25fd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5653337d0e00 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f516f25fd80 4 rocksdb: DB pointer 0x5653337e0000 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f5165029640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-08T23:37:43.803 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f5165029640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: ** DB Stats ** 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: ** Compaction Stats [default] ** 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: L0 4/0 3.39 MB 1.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 86.8 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Sum 4/0 3.39 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 86.8 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 86.8 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: ** Compaction Stats [default] ** 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 86.8 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: AddFile(Total Files): cumulative 0, interval 0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: AddFile(Keys): cumulative 0, interval 0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Cumulative compaction: 0.00 GB write, 13.78 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Interval compaction: 0.00 GB write, 13.78 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Block cache BinnedLRUCache@0x5653337ce9b0#7 capacity: 512.00 MB usage: 12.88 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1e-05 secs_since: 0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: Block cache entry stats(count,size,portion): FilterBlock(4,5.11 KB,0.000974536%) IndexBlock(4,7.77 KB,0.00148118%) Misc(1,0.00 KB,0%) 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: ** File Read Latency Histogram By Level [default] ** 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f516d039640 4 rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 3] Compacting 4@0 files to L6, score 1.00 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f516d039640 4 rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 2 Base level 0, inputs: [23(195KB) 18(3203KB) 13(70KB) 8(1900B)] 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f516d039640 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013063447653, "job": 3, "event": "compaction_started", "compaction_reason": "LevelL0FilesNum", "files_L0": [23, 18, 13, 8], "score": 1, "input_data_size": 3554712, "oldest_snapshot_seqno": -1} 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.444+0000 7f516f25fd80 10 CompressorRegistry(0x565333797418) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.448+0000 7f516602b640 2 Event(0x5653337c9948 nevent=5000 time_id=1).set_owner center_id=0 owner=139987580532288 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.448+0000 7f516602b640 10 stack operator() starting 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.448+0000 7f516682c640 2 Event(0x5653337c9dc8 nevent=5000 time_id=1).set_owner center_id=1 owner=139987588924992 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.448+0000 7f516c037640 2 Event(0x565333880988 nevent=5000 time_id=1).set_owner center_id=2 owner=139987681244736 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.448+0000 7f516c037640 10 stack operator() starting 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.448+0000 7f516682c640 10 stack operator() starting 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.448+0000 7f516f25fd80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] at bind addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.448+0000 7f516f25fd80 10 CompressorRegistry(0x565333797d18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.452+0000 7f516f25fd80 1 mon.a@-1(???) e2 preinit fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.452+0000 7f516f25fd80 0 mon.a@-1(???).mds e1 new map 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.452+0000 7f516f25fd80 0 mon.a@-1(???).mds e1 print_map 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: e1 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: btime 2026-03-08T23:34:14:089156+0000 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: legacy client fscid: -1 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: No filesystems configured 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.452+0000 7f516f25fd80 0 mon.a@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.452+0000 7f516f25fd80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.452+0000 7f516f25fd80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.452+0000 7f516f25fd80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:43.804 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:43 vm02 bash[37576]: debug 2026-03-08T23:37:43.452+0000 7f516f25fd80 1 mon.a@-1(???).paxosservice(auth 1..9) refresh upgraded, format 0 -> 3 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:42.880284+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:42.880284+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:42.880424+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:42.880424+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.866910+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.866910+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.870126+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.870126+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874784+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874784+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874802+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874802+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874812+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874812+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874820+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874820+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874829+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874829+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874837+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874837+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:45.202 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874845+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874845+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874854+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.874854+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.875185+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.875185+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.875212+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.875212+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.875885+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.875885+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.875975+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: cluster 2026-03-08T23:37:44.875975+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.882446+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.882446+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.887795+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.887795+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.888711+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.888711+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.889444+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.889444+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.890201+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:45.203 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:44 vm05 bash[30512]: audit 2026-03-08T23:37:44.890201+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:45.297 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:42.880284+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:42.880284+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:42.880424+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:42.880424+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.866910+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.866910+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.870126+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.870126+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874784+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874784+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874802+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874802+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874812+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874812+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874820+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874820+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874829+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874829+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874837+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874837+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874845+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874845+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874854+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.874854+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.875185+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.875185+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.875212+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.875212+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.875885+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.875885+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.875975+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: cluster 2026-03-08T23:37:44.875975+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.882446+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.882446+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14150 ' entity='' 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.887795+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.887795+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14150 ' entity='mgr.a' 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.888711+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.888711+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.889444+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.889444+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.890201+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:45.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:44 vm02 bash[37576]: audit 2026-03-08T23:37:44.890201+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:45.467 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:45.467 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:45.467 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: Stopping Ceph mon.b for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:37:45.467 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:45.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30512]: debug 2026-03-08T23:37:45.465+0000 7f8efcd7a640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-08T23:37:45.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30512]: debug 2026-03-08T23:37:45.465+0000 7f8efcd7a640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-08T23:37:45.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30862]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mon-b 2026-03-08T23:37:45.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.b.service: Deactivated successfully. 2026-03-08T23:37:45.749 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: Stopped Ceph mon.b for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: Started Ceph mon.b for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.945+0000 7f41fa9e9d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.945+0000 7f41fa9e9d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.945+0000 7f41fa9e9d80 0 pidfile_write: ignore empty --pid-file 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 0 load: jerasure load: lrc 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Git sha 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: DB SUMMARY 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: DB Session ID: RJ1YEWQIY94HQQ1KXTOB 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: CURRENT file: CURRENT 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: MANIFEST file: MANIFEST-000015 size: 282 Bytes 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000014.log size: 337699 ; 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.error_if_exists: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.create_if_missing: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.env: 0x55bcc9cfedc0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.info_log: 0x55bd05a02700 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.statistics: (nil) 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.use_fsync: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.db_log_dir: 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.wal_dir: 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-08T23:37:46.143 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.write_buffer_manager: 0x55bd05a07900 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.unordered_write: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.row_cache: None 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.wal_filter: None 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.two_write_queues: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.wal_compression: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.atomic_flush: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_open_files: -1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Compression algorithms supported: 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: kZSTD supported: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: kXpressCompression supported: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: kZlibCompression supported: 1 2026-03-08T23:37:46.144 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.merge_operator: 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_filter: None 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55bd05a02640) 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: cache_index_and_filter_blocks: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: pin_top_level_index_and_filter: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: index_type: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: data_block_index_type: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: index_shortening: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: data_block_hash_table_util_ratio: 0.750000 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: checksum: 4 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: no_block_cache: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: block_cache: 0x55bd05a29350 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: block_cache_name: BinnedLRUCache 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: block_cache_options: 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: capacity : 536870912 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: num_shard_bits : 4 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: strict_capacity_limit : 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: high_pri_pool_ratio: 0.000 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: block_cache_compressed: (nil) 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: persistent_cache: (nil) 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: block_size: 4096 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: block_size_deviation: 10 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: block_restart_interval: 16 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: index_block_restart_interval: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: metadata_block_size: 4096 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: partition_filters: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: use_delta_encoding: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: filter_policy: bloomfilter 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: whole_key_filtering: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: verify_compression: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: read_amp_bytes_per_bit: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: format_version: 5 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: enable_index_compression: 1 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: block_align: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: max_auto_readahead_size: 262144 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: prepopulate_block_cache: 0 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: initial_auto_readahead_size: 8192 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: num_file_reads_for_auto_readahead: 2 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression: NoCompression 2026-03-08T23:37:46.145 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-08T23:37:46.145 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.num_levels: 7 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-08T23:37:46.146 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.bloom_locality: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.ttl: 2592000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.enable_blob_files: false 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.min_blob_size: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 3242, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 7ce5a3f0-68c5-45fd-bad2-ce95fe1c5ad6 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013065953617, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.949+0000 7f41fa9e9d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.957+0000 7f41fa9e9d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013065960232, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 202506, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 3254, "largest_seqno": 3465, "table_properties": {"data_size": 200416, "index_size": 737, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 325, "raw_key_size": 2737, "raw_average_key_size": 24, "raw_value_size": 197880, "raw_average_value_size": 1782, "num_data_blocks": 32, "num_entries": 111, "num_filter_entries": 111, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773013065, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "7ce5a3f0-68c5-45fd-bad2-ce95fe1c5ad6", "db_session_id": "RJ1YEWQIY94HQQ1KXTOB", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.957+0000 7f41fa9e9d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773013065960286, "job": 1, "event": "recovery_finished"} 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.957+0000 7f41fa9e9d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55bd05a2ae00 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 4 rocksdb: DB pointer 0x55bd05b44000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41f07b3640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41f07b3640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: ** DB Stats ** 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: ** Compaction Stats [default] ** 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: L0 3/0 3.43 MB 0.8 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 117.5 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Sum 3/0 3.43 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 117.5 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 117.5 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: ** Compaction Stats [default] ** 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 117.5 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-08T23:37:46.147 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: AddFile(Total Files): cumulative 0, interval 0 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: AddFile(Keys): cumulative 0, interval 0 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Cumulative compaction: 0.00 GB write, 17.86 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Interval compaction: 0.00 GB write, 17.86 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Block cache BinnedLRUCache@0x55bd05a29350#7 capacity: 512.00 MB usage: 12.03 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 7e-06 secs_since: 0 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: Block cache entry stats(count,size,portion): FilterBlock(3,4.52 KB,0.000861287%) IndexBlock(3,7.52 KB,0.00143349%) Misc(1,0.00 KB,0%) 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: ** File Read Latency Histogram By Level [default] ** 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 10 CompressorRegistry(0x55bd059f1418) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41f17b5640 2 Event(0x55bd05a23dc8 nevent=5000 time_id=1).set_owner center_id=0 owner=139921200993856 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41f17b5640 10 stack operator() starting 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41f77c1640 2 Event(0x55bd068382c8 nevent=5000 time_id=1).set_owner center_id=2 owner=139921301706304 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41f77c1640 10 stack operator() starting 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41f1fb6640 2 Event(0x55bd06838088 nevent=5000 time_id=1).set_owner center_id=1 owner=139921209386560 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41f1fb6640 10 stack operator() starting 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 0 starting mon.b rank 1 at public addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] at bind addrs [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 10 CompressorRegistry(0x55bd059f1d18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 1 mon.b@-1(???) e2 preinit fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 0 mon.b@-1(???).mds e1 new map 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 0 mon.b@-1(???).mds e1 print_map 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: e1 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: btime 2026-03-08T23:34:14:089156+0000 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: legacy client fscid: -1 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: No filesystems configured 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 0 mon.b@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:45 vm05 bash[30991]: debug 2026-03-08T23:37:45.961+0000 7f41fa9e9d80 1 mon.b@-1(???).paxosservice(auth 1..10) refresh upgraded, format 0 -> 3 2026-03-08T23:37:46.148 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:37:45 vm05 systemd[1]: /etc/systemd/system/ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-08T23:37:47.440 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.047568+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.047568+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.051707+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.051707+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.053920+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.053920+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059134+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059134+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059143+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059143+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059147+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059147+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059150+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059150+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059153+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059153+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059156+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059156+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059159+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059159+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059162+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059162+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059462+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059462+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059473+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059473+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059643+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059643+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059713+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: cluster 2026-03-08T23:37:46.059713+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: audit 2026-03-08T23:37:46.066014+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: audit 2026-03-08T23:37:46.066014+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: audit 2026-03-08T23:37:46.070899+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: audit 2026-03-08T23:37:46.070899+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: audit 2026-03-08T23:37:46.072024+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:47.500 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:47 vm05 bash[30991]: audit 2026-03-08T23:37:46.072024+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:47.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.047568+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:37:47.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.047568+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-08T23:37:47.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.051707+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.051707+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.053920+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.053920+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059134+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059134+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059143+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059143+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059147+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059147+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-08T23:35:04.807804+0000 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059150+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059150+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-08T23:34:12.859532+0000 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059153+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059153+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059156+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059156+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059159+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059159+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059162+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059162+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.b 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059462+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059462+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059473+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059473+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059643+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059643+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059713+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: cluster 2026-03-08T23:37:46.059713+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: audit 2026-03-08T23:37:46.066014+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: audit 2026-03-08T23:37:46.066014+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: audit 2026-03-08T23:37:46.070899+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: audit 2026-03-08T23:37:46.070899+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: audit 2026-03-08T23:37:46.072024+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:47.548 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:47 vm02 bash[37576]: audit 2026-03-08T23:37:46.072024+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:37:48.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:48 vm05 bash[30991]: cluster 2026-03-08T23:37:46.584748+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:48.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:48 vm05 bash[30991]: cluster 2026-03-08T23:37:46.584748+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:48.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:48 vm02 bash[37576]: cluster 2026-03-08T23:37:46.584748+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:48.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:48 vm02 bash[37576]: cluster 2026-03-08T23:37:46.584748+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:50.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:50 vm05 bash[30991]: cluster 2026-03-08T23:37:48.585051+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:50.499 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:50 vm05 bash[30991]: cluster 2026-03-08T23:37:48.585051+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:50.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:50 vm02 bash[37576]: cluster 2026-03-08T23:37:48.585051+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:50.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:50 vm02 bash[37576]: cluster 2026-03-08T23:37:48.585051+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:51.464 INFO:teuthology.orchestra.run.vm02.stderr:Using ceph image with id '654f31e6858e' and tag 'e911bdebe5c8faa3800735d1568fcdca65db60df' created on 2026-02-25 18:57:17 +0000 UTC 2026-03-08T23:37:51.464 INFO:teuthology.orchestra.run.vm02.stderr:quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-08T23:37:51.752 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: cluster 2026-03-08T23:37:50.585244+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:51.752 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: cluster 2026-03-08T23:37:50.585244+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:51.752 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:50.632682+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.752 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:50.632682+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.752 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:50.636142+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.752 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:50.636142+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.752 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.022689+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.022689+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.025840+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.025840+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.026889+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.026889+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.027613+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.027613+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.030712+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.753 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:51 vm02 bash[37576]: audit 2026-03-08T23:37:51.030712+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:51.791 INFO:teuthology.orchestra.run.vm02.stderr:+ FSID=32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:37:51.791 INFO:teuthology.orchestra.run.vm02.stderr:+ sleep 60 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: cluster 2026-03-08T23:37:50.585244+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: cluster 2026-03-08T23:37:50.585244+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:50.632682+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:50.632682+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:50.636142+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:50.636142+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.022689+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.022689+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.025840+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.025840+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.026889+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.026889+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.027613+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.027613+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.030712+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:51 vm05 bash[30991]: audit 2026-03-08T23:37:51.030712+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' 2026-03-08T23:37:52.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:52 vm02 bash[37576]: audit 2026-03-08T23:37:51.743688+0000 mon.a (mon.0) 43 : audit [DBG] from='client.? 192.168.123.102:0/4034663879' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-08T23:37:52.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:52 vm02 bash[37576]: audit 2026-03-08T23:37:51.743688+0000 mon.a (mon.0) 43 : audit [DBG] from='client.? 192.168.123.102:0/4034663879' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-08T23:37:52.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:52 vm05 bash[30991]: audit 2026-03-08T23:37:51.743688+0000 mon.a (mon.0) 43 : audit [DBG] from='client.? 192.168.123.102:0/4034663879' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-08T23:37:52.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:52 vm05 bash[30991]: audit 2026-03-08T23:37:51.743688+0000 mon.a (mon.0) 43 : audit [DBG] from='client.? 192.168.123.102:0/4034663879' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-08T23:37:53.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:53 vm05 bash[30991]: cluster 2026-03-08T23:37:52.585475+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:54.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:53 vm05 bash[30991]: cluster 2026-03-08T23:37:52.585475+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:54.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:53 vm02 bash[37576]: cluster 2026-03-08T23:37:52.585475+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:54.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:53 vm02 bash[37576]: cluster 2026-03-08T23:37:52.585475+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:55.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:55 vm05 bash[30991]: cluster 2026-03-08T23:37:54.585742+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:56.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:55 vm05 bash[30991]: cluster 2026-03-08T23:37:54.585742+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:55 vm02 bash[37576]: cluster 2026-03-08T23:37:54.585742+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:56.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:55 vm02 bash[37576]: cluster 2026-03-08T23:37:54.585742+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:57.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:57 vm05 bash[30991]: cluster 2026-03-08T23:37:56.585998+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:58.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:57 vm05 bash[30991]: cluster 2026-03-08T23:37:56.585998+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:58.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:57 vm02 bash[37576]: cluster 2026-03-08T23:37:56.585998+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:58.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:57 vm02 bash[37576]: cluster 2026-03-08T23:37:56.585998+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:59.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:59 vm05 bash[30991]: cluster 2026-03-08T23:37:58.586206+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:37:59.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:37:59 vm05 bash[30991]: cluster 2026-03-08T23:37:58.586206+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:00.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:59 vm02 bash[37576]: cluster 2026-03-08T23:37:58.586206+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:00.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:37:59 vm02 bash[37576]: cluster 2026-03-08T23:37:58.586206+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:01.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:01 vm05 bash[30991]: cluster 2026-03-08T23:38:00.586403+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:02.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:01 vm05 bash[30991]: cluster 2026-03-08T23:38:00.586403+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:02.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:01 vm02 bash[37576]: cluster 2026-03-08T23:38:00.586403+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:02.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:01 vm02 bash[37576]: cluster 2026-03-08T23:38:00.586403+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:03.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:03 vm05 bash[30991]: cluster 2026-03-08T23:38:02.586665+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:03.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:03 vm05 bash[30991]: cluster 2026-03-08T23:38:02.586665+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:04.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:03 vm02 bash[37576]: cluster 2026-03-08T23:38:02.586665+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:04.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:03 vm02 bash[37576]: cluster 2026-03-08T23:38:02.586665+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:05.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:05 vm05 bash[30991]: cluster 2026-03-08T23:38:04.586857+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:06.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:05 vm05 bash[30991]: cluster 2026-03-08T23:38:04.586857+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:06.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:05 vm02 bash[37576]: cluster 2026-03-08T23:38:04.586857+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:06.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:05 vm02 bash[37576]: cluster 2026-03-08T23:38:04.586857+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:07.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:07 vm05 bash[30991]: cluster 2026-03-08T23:38:06.587092+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:08.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:07 vm05 bash[30991]: cluster 2026-03-08T23:38:06.587092+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:08.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:07 vm02 bash[37576]: cluster 2026-03-08T23:38:06.587092+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:08.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:07 vm02 bash[37576]: cluster 2026-03-08T23:38:06.587092+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:09.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:09 vm05 bash[30991]: cluster 2026-03-08T23:38:08.587334+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:09.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:09 vm05 bash[30991]: cluster 2026-03-08T23:38:08.587334+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:10.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:09 vm02 bash[37576]: cluster 2026-03-08T23:38:08.587334+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:10.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:09 vm02 bash[37576]: cluster 2026-03-08T23:38:08.587334+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:11.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:11 vm05 bash[30991]: cluster 2026-03-08T23:38:10.587549+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:12.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:11 vm05 bash[30991]: cluster 2026-03-08T23:38:10.587549+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:12.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:11 vm02 bash[37576]: cluster 2026-03-08T23:38:10.587549+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:12.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:11 vm02 bash[37576]: cluster 2026-03-08T23:38:10.587549+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:13.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:13 vm05 bash[30991]: cluster 2026-03-08T23:38:12.587803+0000 mgr.a (mgr.14150) 171 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:14.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:13 vm05 bash[30991]: cluster 2026-03-08T23:38:12.587803+0000 mgr.a (mgr.14150) 171 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:14.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:13 vm02 bash[37576]: cluster 2026-03-08T23:38:12.587803+0000 mgr.a (mgr.14150) 171 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:14.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:13 vm02 bash[37576]: cluster 2026-03-08T23:38:12.587803+0000 mgr.a (mgr.14150) 171 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:15.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:15 vm05 bash[30991]: cluster 2026-03-08T23:38:14.588040+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:16.000 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:15 vm05 bash[30991]: cluster 2026-03-08T23:38:14.588040+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:16.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:15 vm02 bash[37576]: cluster 2026-03-08T23:38:14.588040+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:16.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:15 vm02 bash[37576]: cluster 2026-03-08T23:38:14.588040+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:17.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:17 vm05 bash[30991]: cluster 2026-03-08T23:38:16.588271+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:17.999 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:17 vm05 bash[30991]: cluster 2026-03-08T23:38:16.588271+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:18.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:17 vm02 bash[37576]: cluster 2026-03-08T23:38:16.588271+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:18.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:17 vm02 bash[37576]: cluster 2026-03-08T23:38:16.588271+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:19 vm02 bash[37576]: cluster 2026-03-08T23:38:18.588479+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:20.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:19 vm02 bash[37576]: cluster 2026-03-08T23:38:18.588479+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:20.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:19 vm05 bash[30991]: cluster 2026-03-08T23:38:18.588479+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:20.250 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:19 vm05 bash[30991]: cluster 2026-03-08T23:38:18.588479+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:22.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:21 vm02 bash[37576]: cluster 2026-03-08T23:38:20.588699+0000 mgr.a (mgr.14150) 175 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:22.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:21 vm02 bash[37576]: cluster 2026-03-08T23:38:20.588699+0000 mgr.a (mgr.14150) 175 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:22.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:21 vm05 bash[30991]: cluster 2026-03-08T23:38:20.588699+0000 mgr.a (mgr.14150) 175 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:22.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:21 vm05 bash[30991]: cluster 2026-03-08T23:38:20.588699+0000 mgr.a (mgr.14150) 175 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:23 vm02 bash[37576]: cluster 2026-03-08T23:38:22.588947+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:24.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:23 vm02 bash[37576]: cluster 2026-03-08T23:38:22.588947+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:24.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:23 vm05 bash[30991]: cluster 2026-03-08T23:38:22.588947+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:24.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:23 vm05 bash[30991]: cluster 2026-03-08T23:38:22.588947+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:26.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:25 vm02 bash[37576]: cluster 2026-03-08T23:38:24.589204+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:26.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:25 vm02 bash[37576]: cluster 2026-03-08T23:38:24.589204+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:26.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:25 vm05 bash[30991]: cluster 2026-03-08T23:38:24.589204+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:26.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:25 vm05 bash[30991]: cluster 2026-03-08T23:38:24.589204+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:28.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:27 vm02 bash[37576]: cluster 2026-03-08T23:38:26.589412+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:28.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:27 vm02 bash[37576]: cluster 2026-03-08T23:38:26.589412+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:28.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:27 vm05 bash[30991]: cluster 2026-03-08T23:38:26.589412+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:28.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:27 vm05 bash[30991]: cluster 2026-03-08T23:38:26.589412+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:30.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:29 vm02 bash[37576]: cluster 2026-03-08T23:38:28.589689+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:30.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:29 vm02 bash[37576]: cluster 2026-03-08T23:38:28.589689+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:30.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:29 vm05 bash[30991]: cluster 2026-03-08T23:38:28.589689+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:30.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:29 vm05 bash[30991]: cluster 2026-03-08T23:38:28.589689+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:32.047 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:31 vm02 bash[37576]: cluster 2026-03-08T23:38:30.589910+0000 mgr.a (mgr.14150) 180 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:32.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:31 vm02 bash[37576]: cluster 2026-03-08T23:38:30.589910+0000 mgr.a (mgr.14150) 180 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:32.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:31 vm05 bash[30991]: cluster 2026-03-08T23:38:30.589910+0000 mgr.a (mgr.14150) 180 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:32.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:31 vm05 bash[30991]: cluster 2026-03-08T23:38:30.589910+0000 mgr.a (mgr.14150) 180 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:34.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:33 vm05 bash[30991]: cluster 2026-03-08T23:38:32.590154+0000 mgr.a (mgr.14150) 181 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:34.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:33 vm05 bash[30991]: cluster 2026-03-08T23:38:32.590154+0000 mgr.a (mgr.14150) 181 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:34.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:33 vm02 bash[37576]: cluster 2026-03-08T23:38:32.590154+0000 mgr.a (mgr.14150) 181 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:34.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:33 vm02 bash[37576]: cluster 2026-03-08T23:38:32.590154+0000 mgr.a (mgr.14150) 181 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:36.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:35 vm05 bash[30991]: cluster 2026-03-08T23:38:34.590345+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:36.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:35 vm05 bash[30991]: cluster 2026-03-08T23:38:34.590345+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:36.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:35 vm02 bash[37576]: cluster 2026-03-08T23:38:34.590345+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:36.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:35 vm02 bash[37576]: cluster 2026-03-08T23:38:34.590345+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:38.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:37 vm05 bash[30991]: cluster 2026-03-08T23:38:36.590574+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:38.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:37 vm05 bash[30991]: cluster 2026-03-08T23:38:36.590574+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:38.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:37 vm02 bash[37576]: cluster 2026-03-08T23:38:36.590574+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:38.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:37 vm02 bash[37576]: cluster 2026-03-08T23:38:36.590574+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:40.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:39 vm05 bash[30991]: cluster 2026-03-08T23:38:38.590890+0000 mgr.a (mgr.14150) 184 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:40.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:39 vm05 bash[30991]: cluster 2026-03-08T23:38:38.590890+0000 mgr.a (mgr.14150) 184 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:40.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:39 vm02 bash[37576]: cluster 2026-03-08T23:38:38.590890+0000 mgr.a (mgr.14150) 184 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:40.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:39 vm02 bash[37576]: cluster 2026-03-08T23:38:38.590890+0000 mgr.a (mgr.14150) 184 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:42.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:41 vm05 bash[30991]: cluster 2026-03-08T23:38:40.591096+0000 mgr.a (mgr.14150) 185 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:42.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:41 vm05 bash[30991]: cluster 2026-03-08T23:38:40.591096+0000 mgr.a (mgr.14150) 185 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:42.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:41 vm02 bash[37576]: cluster 2026-03-08T23:38:40.591096+0000 mgr.a (mgr.14150) 185 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:42.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:41 vm02 bash[37576]: cluster 2026-03-08T23:38:40.591096+0000 mgr.a (mgr.14150) 185 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:44.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:43 vm05 bash[30991]: cluster 2026-03-08T23:38:42.591314+0000 mgr.a (mgr.14150) 186 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:44.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:43 vm05 bash[30991]: cluster 2026-03-08T23:38:42.591314+0000 mgr.a (mgr.14150) 186 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:44.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:43 vm02 bash[37576]: cluster 2026-03-08T23:38:42.591314+0000 mgr.a (mgr.14150) 186 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:44.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:43 vm02 bash[37576]: cluster 2026-03-08T23:38:42.591314+0000 mgr.a (mgr.14150) 186 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:46.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:45 vm05 bash[30991]: cluster 2026-03-08T23:38:44.591584+0000 mgr.a (mgr.14150) 187 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:46.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:45 vm05 bash[30991]: cluster 2026-03-08T23:38:44.591584+0000 mgr.a (mgr.14150) 187 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:46.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:45 vm02 bash[37576]: cluster 2026-03-08T23:38:44.591584+0000 mgr.a (mgr.14150) 187 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:46.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:45 vm02 bash[37576]: cluster 2026-03-08T23:38:44.591584+0000 mgr.a (mgr.14150) 187 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:48.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:47 vm05 bash[30991]: cluster 2026-03-08T23:38:46.591888+0000 mgr.a (mgr.14150) 188 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:48.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:47 vm05 bash[30991]: cluster 2026-03-08T23:38:46.591888+0000 mgr.a (mgr.14150) 188 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:48.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:47 vm02 bash[37576]: cluster 2026-03-08T23:38:46.591888+0000 mgr.a (mgr.14150) 188 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:48.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:47 vm02 bash[37576]: cluster 2026-03-08T23:38:46.591888+0000 mgr.a (mgr.14150) 188 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:50.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:49 vm05 bash[30991]: cluster 2026-03-08T23:38:48.592084+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:50.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:49 vm05 bash[30991]: cluster 2026-03-08T23:38:48.592084+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:50.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:49 vm02 bash[37576]: cluster 2026-03-08T23:38:48.592084+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:50.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:49 vm02 bash[37576]: cluster 2026-03-08T23:38:48.592084+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:51.793 INFO:teuthology.orchestra.run.vm02.stderr:+ grep '\-\-cpus=2' /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/unit.run 2026-03-08T23:38:51.793 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mon-a --pids-limit=0 --cpus=2 -e CONTAINER_IMAGE=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc -e NODE_NAME=vm02 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109:/var/run/ceph:z -v /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109:/var/log/ceph:z -v /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config:/etc/ceph/ceph.conf:z quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true '--default-log-stderr-prefix=debug ' --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 2026-03-08T23:38:51.793 INFO:teuthology.orchestra.run.vm02.stderr:+ grep '\-\-debug_ms 10' /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/unit.run 2026-03-08T23:38:51.794 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mon-a --pids-limit=0 --cpus=2 -e CONTAINER_IMAGE=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc -e NODE_NAME=vm02 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109:/var/run/ceph:z -v /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109:/var/log/ceph:z -v /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config:/etc/ceph/ceph.conf:z quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true '--default-log-stderr-prefix=debug ' --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 2026-03-08T23:38:51.794 INFO:teuthology.orchestra.run.vm02.stderr:+ grep testing_custom_containers /etc/cephadm_testing/testing.txt 2026-03-08T23:38:51.795 INFO:teuthology.orchestra.run.vm02.stdout:testing_custom_containers 2026-03-08T23:38:51.795 INFO:teuthology.orchestra.run.vm02.stderr:++ find /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109 -maxdepth 1 -type d -name 'container.bar.*' 2026-03-08T23:38:51.796 INFO:teuthology.orchestra.run.vm02.stderr:+ dir=/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/container.bar.vm02 2026-03-08T23:38:51.796 INFO:teuthology.orchestra.run.vm02.stderr:+ test -n /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/container.bar.vm02 2026-03-08T23:38:51.796 INFO:teuthology.orchestra.run.vm02.stderr:+ grep ok /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/container.bar.vm02/data/primary.txt 2026-03-08T23:38:51.796 INFO:teuthology.orchestra.run.vm02.stdout:ok 2026-03-08T23:38:51.797 INFO:teuthology.orchestra.run.vm02.stderr:+ grep from=centos /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/container.bar.vm02/data/from.txt 2026-03-08T23:38:51.797 INFO:teuthology.orchestra.run.vm02.stdout:from=centos 2026-03-08T23:38:51.797 INFO:teuthology.orchestra.run.vm02.stderr:+ test -s /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/container.bar.vm02/data/presized.dat 2026-03-08T23:38:51.798 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-08T23:38:51.800 INFO:tasks.cephadm:Teardown begin 2026-03-08T23:38:51.800 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-08T23:38:51.849 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-08T23:38:51.861 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-08T23:38:51.861 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 -- ceph mgr module disable cephadm 2026-03-08T23:38:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:51 vm02 bash[37576]: cluster 2026-03-08T23:38:50.592263+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:51 vm02 bash[37576]: cluster 2026-03-08T23:38:50.592263+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:51 vm02 bash[37576]: audit 2026-03-08T23:38:51.042441+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:38:52.048 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:51 vm02 bash[37576]: audit 2026-03-08T23:38:51.042441+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:38:52.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:51 vm05 bash[30991]: cluster 2026-03-08T23:38:50.592263+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:52.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:51 vm05 bash[30991]: cluster 2026-03-08T23:38:50.592263+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:52.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:51 vm05 bash[30991]: audit 2026-03-08T23:38:51.042441+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:38:52.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:51 vm05 bash[30991]: audit 2026-03-08T23:38:51.042441+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14150 192.168.123.102:0/3980328845' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-08T23:38:54.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:53 vm05 bash[30991]: cluster 2026-03-08T23:38:52.592462+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:54.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:53 vm05 bash[30991]: cluster 2026-03-08T23:38:52.592462+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:54.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:53 vm02 bash[37576]: cluster 2026-03-08T23:38:52.592462+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:54.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:53 vm02 bash[37576]: cluster 2026-03-08T23:38:52.592462+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:55.818 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/mon.a/config 2026-03-08T23:38:55.970 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-08T23:38:55.963+0000 7ff6f4cc6640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-08T23:38:55.970 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-08T23:38:55.963+0000 7ff6f4cc6640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-08T23:38:55.974 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-08T23:38:55.971+0000 7ff6f4cc6640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-08T23:38:55.974 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-08T23:38:55.971+0000 7ff6f4cc6640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-08T23:38:55.974 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-08T23:38:55.971+0000 7ff6f4cc6640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-08T23:38:55.974 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-08T23:38:55.971+0000 7ff6f4cc6640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-08T23:38:55.974 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-08T23:38:55.971+0000 7ff6f4cc6640 -1 monclient: keyring not found 2026-03-08T23:38:55.975 INFO:teuthology.orchestra.run.vm02.stderr:[errno 21] error connecting to the cluster 2026-03-08T23:38:56.055 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:38:56.055 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-08T23:38:56.055 DEBUG:teuthology.orchestra.run.vm02:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-08T23:38:56.059 DEBUG:teuthology.orchestra.run.vm05:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-08T23:38:56.062 INFO:tasks.cephadm:Stopping all daemons... 2026-03-08T23:38:56.063 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-08T23:38:56.063 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a 2026-03-08T23:38:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:55 vm02 bash[37576]: cluster 2026-03-08T23:38:54.592712+0000 mgr.a (mgr.14150) 192 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:55 vm02 bash[37576]: cluster 2026-03-08T23:38:54.592712+0000 mgr.a (mgr.14150) 192 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:56.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:55 vm05 bash[30991]: cluster 2026-03-08T23:38:54.592712+0000 mgr.a (mgr.14150) 192 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:56.249 INFO:journalctl@ceph.mon.b.vm05.stdout:Mar 08 23:38:55 vm05 bash[30991]: cluster 2026-03-08T23:38:54.592712+0000 mgr.a (mgr.14150) 192 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-08T23:38:56.301 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service' 2026-03-08T23:38:56.347 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:56 vm02 systemd[1]: Stopping Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:38:56.347 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:56 vm02 bash[37576]: debug 2026-03-08T23:38:56.151+0000 7f516d83a640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 (PID: 1) UID: 0 2026-03-08T23:38:56.347 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:56 vm02 bash[37576]: debug 2026-03-08T23:38:56.151+0000 7f516d83a640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-08T23:38:56.347 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:56 vm02 bash[38196]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mon-a 2026-03-08T23:38:56.347 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:56 vm02 systemd[1]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.a.service: Deactivated successfully. 2026-03-08T23:38:56.347 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 08 23:38:56 vm02 systemd[1]: Stopped Ceph mon.a for 32c51f16-1b47-11f1-bdef-c3bc34602109. 2026-03-08T23:38:56.355 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-08T23:38:56.355 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-08T23:38:56.355 INFO:tasks.cephadm.mon.b:Stopping mon.b... 2026-03-08T23:38:56.355 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.b 2026-03-08T23:38:56.604 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mon.b.service' 2026-03-08T23:38:56.614 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-08T23:38:56.614 INFO:tasks.cephadm.mon.b:Stopped mon.b 2026-03-08T23:38:56.614 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-08T23:38:56.614 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.a 2026-03-08T23:38:56.748 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.a.service' 2026-03-08T23:38:56.757 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-08T23:38:56.757 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-08T23:38:56.757 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-08T23:38:56.757 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.b 2026-03-08T23:38:56.875 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:38:56 vm05 systemd[1]: Stopping Ceph mgr.b for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:38:56.875 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:38:56 vm05 bash[31473]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-mgr-b 2026-03-08T23:38:56.875 INFO:journalctl@ceph.mgr.b.vm05.stdout:Mar 08 23:38:56 vm05 systemd[1]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.b.service: Main process exited, code=exited, status=143/n/a 2026-03-08T23:38:56.876 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@mgr.b.service' 2026-03-08T23:38:56.885 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-08T23:38:56.886 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-08T23:38:56.886 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-08T23:38:56.886 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@osd.0 2026-03-08T23:38:57.298 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:38:56 vm02 systemd[1]: Stopping Ceph osd.0 for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:38:57.298 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:38:56 vm02 bash[30269]: debug 2026-03-08T23:38:56.923+0000 7fd217eb9640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-08T23:38:57.298 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:38:56 vm02 bash[30269]: debug 2026-03-08T23:38:56.923+0000 7fd217eb9640 -1 osd.0 14 *** Got signal Terminated *** 2026-03-08T23:38:57.298 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:38:56 vm02 bash[30269]: debug 2026-03-08T23:38:56.923+0000 7fd217eb9640 -1 osd.0 14 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-08T23:39:02.213 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 08 23:39:01 vm02 bash[38378]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-osd-0 2026-03-08T23:39:02.253 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@osd.0.service' 2026-03-08T23:39:02.275 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-08T23:39:02.275 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-08T23:39:02.275 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-08T23:39:02.275 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@osd.1 2026-03-08T23:39:02.749 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:39:02 vm05 systemd[1]: Stopping Ceph osd.1 for 32c51f16-1b47-11f1-bdef-c3bc34602109... 2026-03-08T23:39:02.749 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:39:02 vm05 bash[25967]: debug 2026-03-08T23:39:02.314+0000 7fe71148b640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-08T23:39:02.749 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:39:02 vm05 bash[25967]: debug 2026-03-08T23:39:02.314+0000 7fe71148b640 -1 osd.1 14 *** Got signal Terminated *** 2026-03-08T23:39:02.749 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:39:02 vm05 bash[25967]: debug 2026-03-08T23:39:02.314+0000 7fe71148b640 -1 osd.1 14 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-08T23:39:07.634 INFO:journalctl@ceph.osd.1.vm05.stdout:Mar 08 23:39:07 vm05 bash[31560]: ceph-32c51f16-1b47-11f1-bdef-c3bc34602109-osd-1 2026-03-08T23:39:07.653 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-32c51f16-1b47-11f1-bdef-c3bc34602109@osd.1.service' 2026-03-08T23:39:07.670 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-08T23:39:07.670 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-08T23:39:07.670 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 --force --keep-logs 2026-03-08T23:39:07.751 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:39:14.364 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 --force --keep-logs 2026-03-08T23:39:14.454 INFO:teuthology.orchestra.run.vm05.stdout:Deleting cluster with fsid: 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:39:21.147 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-08T23:39:21.154 INFO:teuthology.orchestra.run.vm02.stderr:rm: cannot remove '/etc/ceph/ceph.client.admin.keyring': Is a directory 2026-03-08T23:39:21.155 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:39:21.155 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-08T23:39:21.162 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-08T23:39:21.162 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/crash to /archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300/remote/vm02/crash 2026-03-08T23:39:21.162 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/crash -- . 2026-03-08T23:39:21.205 INFO:teuthology.orchestra.run.vm02.stderr:tar: /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/crash: Cannot open: No such file or directory 2026-03-08T23:39:21.205 INFO:teuthology.orchestra.run.vm02.stderr:tar: Error is not recoverable: exiting now 2026-03-08T23:39:21.206 DEBUG:teuthology.misc:Transferring archived files from vm05:/var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/crash to /archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300/remote/vm05/crash 2026-03-08T23:39:21.206 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/crash -- . 2026-03-08T23:39:21.213 INFO:teuthology.orchestra.run.vm05.stderr:tar: /var/lib/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/crash: Cannot open: No such file or directory 2026-03-08T23:39:21.213 INFO:teuthology.orchestra.run.vm05.stderr:tar: Error is not recoverable: exiting now 2026-03-08T23:39:21.213 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-08T23:39:21.213 DEBUG:teuthology.orchestra.run.vm02:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-08T23:39:21.256 INFO:tasks.cephadm:Compressing logs... 2026-03-08T23:39:21.256 DEBUG:teuthology.orchestra.run.vm02:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-08T23:39:21.299 DEBUG:teuthology.orchestra.run.vm05:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-08T23:39:21.305 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-08T23:39:21.305 INFO:teuthology.orchestra.run.vm02.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-08T23:39:21.306 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mgr.a.log 2026-03-08T23:39:21.306 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.log 2026-03-08T23:39:21.307 INFO:teuthology.orchestra.run.vm05.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-08T23:39:21.307 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mgr.a.log: 89.8% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-08T23:39:21.307 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mon.a.log 2026-03-08T23:39:21.307 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-08T23:39:21.308 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.log 2026-03-08T23:39:21.308 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.log: 87.0% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.log.gz 2026-03-08T23:39:21.308 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mon.b.log 2026-03-08T23:39:21.308 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.audit.log 2026-03-08T23:39:21.309 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.log: 87.2% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.log.gz 2026-03-08T23:39:21.309 INFO:teuthology.orchestra.run.vm05.stderr: 88.8% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-08T23:39:21.309 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-osd.1.log 2026-03-08T23:39:21.309 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mgr.b.log 2026-03-08T23:39:21.314 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mon.a.log: gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-volume.log 2026-03-08T23:39:21.315 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.audit.log: 89.7% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.audit.log.gz 2026-03-08T23:39:21.318 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.cephadm.log 2026-03-08T23:39:21.319 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mon.b.log: /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-osd.1.log: 94.1% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-osd.1.log.gz 2026-03-08T23:39:21.319 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.audit.log 2026-03-08T23:39:21.321 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mgr.b.log: 91.1% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mgr.b.log.gz 2026-03-08T23:39:21.322 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-volume.log 2026-03-08T23:39:21.322 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.audit.log: 90.1% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.audit.log.gz 2026-03-08T23:39:21.322 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.cephadm.log 2026-03-08T23:39:21.325 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-volume.log: 95.8% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-volume.log.gz 2026-03-08T23:39:21.326 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.cephadm.log: 76.9% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.cephadm.log.gz 2026-03-08T23:39:21.328 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-volume.log: 95.9% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-volume.log.gz 2026-03-08T23:39:21.329 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-osd.0.log 2026-03-08T23:39:21.329 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.cephadm.log: 79.2% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph.cephadm.log.gz 2026-03-08T23:39:21.339 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-osd.0.log: 89.8% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mgr.a.log.gz 2026-03-08T23:39:21.347 INFO:teuthology.orchestra.run.vm02.stderr: 94.1% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-osd.0.log.gz 2026-03-08T23:39:21.348 INFO:teuthology.orchestra.run.vm05.stderr: 93.5% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mon.b.log.gz 2026-03-08T23:39:21.350 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-08T23:39:21.350 INFO:teuthology.orchestra.run.vm05.stderr:real 0m0.048s 2026-03-08T23:39:21.350 INFO:teuthology.orchestra.run.vm05.stderr:user 0m0.059s 2026-03-08T23:39:21.350 INFO:teuthology.orchestra.run.vm05.stderr:sys 0m0.008s 2026-03-08T23:39:21.417 INFO:teuthology.orchestra.run.vm02.stderr: 91.8% -- replaced with /var/log/ceph/32c51f16-1b47-11f1-bdef-c3bc34602109/ceph-mon.a.log.gz 2026-03-08T23:39:21.417 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-08T23:39:21.417 INFO:teuthology.orchestra.run.vm02.stderr:real 0m0.117s 2026-03-08T23:39:21.417 INFO:teuthology.orchestra.run.vm02.stderr:user 0m0.145s 2026-03-08T23:39:21.417 INFO:teuthology.orchestra.run.vm02.stderr:sys 0m0.013s 2026-03-08T23:39:21.418 INFO:tasks.cephadm:Archiving logs... 2026-03-08T23:39:21.418 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/log/ceph to /archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300/remote/vm02/log 2026-03-08T23:39:21.418 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-08T23:39:21.474 DEBUG:teuthology.misc:Transferring archived files from vm05:/var/log/ceph to /archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300/remote/vm05/log 2026-03-08T23:39:21.475 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-08T23:39:21.487 INFO:tasks.cephadm:Removing cluster... 2026-03-08T23:39:21.488 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 --force 2026-03-08T23:39:21.598 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:39:22.832 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 32c51f16-1b47-11f1-bdef-c3bc34602109 --force 2026-03-08T23:39:22.923 INFO:teuthology.orchestra.run.vm05.stdout:Deleting cluster with fsid: 32c51f16-1b47-11f1-bdef-c3bc34602109 2026-03-08T23:39:24.190 INFO:tasks.cephadm:Removing cephadm ... 2026-03-08T23:39:24.190 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-08T23:39:24.193 DEBUG:teuthology.orchestra.run.vm05:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-08T23:39:24.197 INFO:tasks.cephadm:Teardown complete 2026-03-08T23:39:24.197 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-08T23:39:24.199 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-08T23:39:24.200 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-08T23:39:24.235 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-08T23:39:24.262 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-08T23:39:24.262 DEBUG:teuthology.orchestra.run.vm02:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-08T23:39:24.266 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-08T23:39:24.266 DEBUG:teuthology.orchestra.run.vm05:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-08T23:39:24.318 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:24.329 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:24.500 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:24.500 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:24.527 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:24.527 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:24.688 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:24.688 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:39:24.689 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-08T23:39:24.689 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:24.690 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:24.690 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:39:24.690 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-08T23:39:24.690 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:24.698 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:24.698 INFO:teuthology.orchestra.run.vm05.stdout: ceph* 2026-03-08T23:39:24.704 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:24.705 INFO:teuthology.orchestra.run.vm02.stdout: ceph* 2026-03-08T23:39:24.871 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-08T23:39:24.871 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-08T23:39:24.889 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-08T23:39:24.889 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-08T23:39:24.905 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118605 files and directories currently installed.) 2026-03-08T23:39:24.907 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:24.925 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118605 files and directories currently installed.) 2026-03-08T23:39:24.927 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:26.046 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:26.079 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:26.123 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:26.157 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:26.254 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:26.254 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:26.363 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:26.364 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:26.414 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:26.414 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:39:26.414 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-08T23:39:26.414 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:26.427 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:26.428 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm* cephadm* 2026-03-08T23:39:26.530 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:26.530 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:39:26.530 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-08T23:39:26.530 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:26.542 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:26.543 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm* cephadm* 2026-03-08T23:39:26.608 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-08T23:39:26.608 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 1775 kB disk space will be freed. 2026-03-08T23:39:26.648 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118603 files and directories currently installed.) 2026-03-08T23:39:26.650 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:26.668 INFO:teuthology.orchestra.run.vm05.stdout:Removing cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:26.696 INFO:teuthology.orchestra.run.vm05.stdout:Looking for files to backup/remove ... 2026-03-08T23:39:26.697 INFO:teuthology.orchestra.run.vm05.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-08T23:39:26.699 INFO:teuthology.orchestra.run.vm05.stdout:Removing user `cephadm' ... 2026-03-08T23:39:26.699 INFO:teuthology.orchestra.run.vm05.stdout:Warning: group `nogroup' has no more members. 2026-03-08T23:39:26.709 INFO:teuthology.orchestra.run.vm05.stdout:Done. 2026-03-08T23:39:26.719 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-08T23:39:26.719 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 1775 kB disk space will be freed. 2026-03-08T23:39:26.730 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:26.756 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118603 files and directories currently installed.) 2026-03-08T23:39:26.758 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:26.776 INFO:teuthology.orchestra.run.vm02.stdout:Removing cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:26.804 INFO:teuthology.orchestra.run.vm02.stdout:Looking for files to backup/remove ... 2026-03-08T23:39:26.805 INFO:teuthology.orchestra.run.vm02.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-08T23:39:26.807 INFO:teuthology.orchestra.run.vm02.stdout:Removing user `cephadm' ... 2026-03-08T23:39:26.808 INFO:teuthology.orchestra.run.vm02.stdout:Warning: group `nogroup' has no more members. 2026-03-08T23:39:26.819 INFO:teuthology.orchestra.run.vm02.stdout:Done. 2026-03-08T23:39:26.829 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-08T23:39:26.831 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:26.842 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:26.942 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-08T23:39:26.945 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:27.968 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:28.006 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:28.044 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:28.079 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:28.220 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:28.221 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:28.284 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:28.285 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:28.383 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:28.383 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:39:28.383 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-08T23:39:28.384 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:28.396 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:28.397 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds* 2026-03-08T23:39:28.468 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:28.469 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-08T23:39:28.469 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-08T23:39:28.469 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:28.483 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:28.484 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds* 2026-03-08T23:39:28.585 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-08T23:39:28.585 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 7437 kB disk space will be freed. 2026-03-08T23:39:28.623 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-08T23:39:28.625 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:28.663 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-08T23:39:28.663 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 7437 kB disk space will be freed. 2026-03-08T23:39:28.700 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-08T23:39:28.703 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:29.039 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:29.123 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:29.140 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-08T23:39:29.143 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:29.214 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-08T23:39:29.216 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:30.571 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:30.608 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:30.647 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:30.682 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:30.776 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:30.776 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:30.839 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:30.840 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:30.877 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools python3-cheroot 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-psutil python3-pyinotify 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-threadpoolctl python3-waitress python3-webob python3-websocket 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout: sg3-utils-udev 2026-03-08T23:39:30.878 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:30.886 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:30.886 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-08T23:39:30.886 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents* 2026-03-08T23:39:30.973 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:30.973 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools python3-cheroot 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-psutil python3-pyinotify 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-threadpoolctl python3-waitress python3-webob python3-websocket 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev 2026-03-08T23:39:30.974 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:30.987 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:30.988 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-08T23:39:30.988 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-k8sevents* 2026-03-08T23:39:31.053 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 4 to remove and 10 not upgraded. 2026-03-08T23:39:31.053 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 165 MB disk space will be freed. 2026-03-08T23:39:31.092 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-08T23:39:31.094 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.106 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.135 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.161 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 4 to remove and 10 not upgraded. 2026-03-08T23:39:31.161 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 165 MB disk space will be freed. 2026-03-08T23:39:31.175 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.205 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-08T23:39:31.208 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.220 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.246 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.286 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.696 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-08T23:39:31.699 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:31.806 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-08T23:39:31.808 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:33.219 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:33.252 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:33.278 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:33.311 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:33.453 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:33.454 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:33.480 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:33.481 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:33.647 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:33.647 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:33.647 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:33.648 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:33.663 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:33.664 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-08T23:39:33.675 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:33.675 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:33.675 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:33.676 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:33.691 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:33.692 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-08T23:39:33.851 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-08T23:39:33.851 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 472 MB disk space will be freed. 2026-03-08T23:39:33.874 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-08T23:39:33.875 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 472 MB disk space will be freed. 2026-03-08T23:39:33.899 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-08T23:39:33.902 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:33.914 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-08T23:39:33.916 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:33.967 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:33.977 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:34.399 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:34.440 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:34.814 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:34.878 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:35.262 INFO:teuthology.orchestra.run.vm02.stdout:Removing radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:35.281 INFO:teuthology.orchestra.run.vm05.stdout:Removing radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:35.656 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:35.687 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:35.692 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:35.724 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:36.243 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:36.249 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:36.280 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:39:36.282 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:39:36.358 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117455 files and directories currently installed.) 2026-03-08T23:39:36.360 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:36.363 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117455 files and directories currently installed.) 2026-03-08T23:39:36.365 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:36.960 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:36.988 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:37.396 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:37.399 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:37.827 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:37.832 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:38.248 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:38.262 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:39.726 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:39.762 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:39.875 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:39.909 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:39.931 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:39.932 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:40.038 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:40.038 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:40.039 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:40.046 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:40.047 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse* 2026-03-08T23:39:40.119 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:40.120 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:40.217 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-08T23:39:40.217 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 3673 kB disk space will be freed. 2026-03-08T23:39:40.251 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117443 files and directories currently installed.) 2026-03-08T23:39:40.253 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:40.273 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:40.273 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:40.273 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:40.274 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:40.283 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:40.283 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse* 2026-03-08T23:39:40.455 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-08T23:39:40.455 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 3673 kB disk space will be freed. 2026-03-08T23:39:40.495 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117443 files and directories currently installed.) 2026-03-08T23:39:40.497 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:40.679 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:40.777 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-08T23:39:40.779 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:40.925 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:41.019 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-08T23:39:41.020 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:42.419 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:42.452 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:42.452 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:42.487 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:42.655 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:42.656 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:42.660 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:42.661 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:42.811 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:42.812 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:42.834 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:42.834 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:42.861 INFO:teuthology.orchestra.run.vm02.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-08T23:39:42.861 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:42.861 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:42.861 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:42.862 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:42.862 INFO:teuthology.orchestra.run.vm02.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:42.863 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:42.869 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:42.887 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:42.888 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:42.924 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:43.083 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:43.084 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:43.135 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:43.135 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:43.253 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:43.267 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:43.267 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:43.297 INFO:teuthology.orchestra.run.vm02.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:43.298 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:43.300 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:43.316 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:43.316 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:43.351 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:43.462 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:43.463 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:43.527 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:43.528 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout:Package 'radosgw' is not installed, so not removed 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:43.622 INFO:teuthology.orchestra.run.vm05.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:43.623 INFO:teuthology.orchestra.run.vm05.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:43.623 INFO:teuthology.orchestra.run.vm05.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:43.623 INFO:teuthology.orchestra.run.vm05.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:43.623 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:43.642 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:43.643 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:43.675 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout:Package 'radosgw' is not installed, so not removed 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-08T23:39:43.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-08T23:39:43.679 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-08T23:39:43.679 INFO:teuthology.orchestra.run.vm02.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-08T23:39:43.679 INFO:teuthology.orchestra.run.vm02.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-08T23:39:43.679 INFO:teuthology.orchestra.run.vm02.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-08T23:39:43.679 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:43.692 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:43.692 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:43.725 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:43.861 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:43.861 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:43.891 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:43.892 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:44.009 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:44.009 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:44.009 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:44.009 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-08T23:39:44.010 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:44.024 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:44.024 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-08T23:39:44.065 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:44.065 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:44.065 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:44.065 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet zip 2026-03-08T23:39:44.066 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:44.079 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:44.080 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-08T23:39:44.196 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 3 to remove and 10 not upgraded. 2026-03-08T23:39:44.196 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 2062 kB disk space will be freed. 2026-03-08T23:39:44.230 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-08T23:39:44.232 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:44.250 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:44.251 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 3 to remove and 10 not upgraded. 2026-03-08T23:39:44.251 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 2062 kB disk space will be freed. 2026-03-08T23:39:44.271 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:44.300 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-08T23:39:44.304 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:44.325 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:44.339 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:45.486 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:45.521 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:45.567 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:45.600 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:45.726 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:45.726 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:45.809 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:45.810 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:45.897 INFO:teuthology.orchestra.run.vm05.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-08T23:39:45.897 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:45.897 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:45.897 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:45.898 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:45.899 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:45.899 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-08T23:39:45.899 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:45.920 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:45.920 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:45.954 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:46.008 INFO:teuthology.orchestra.run.vm02.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet zip 2026-03-08T23:39:46.009 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:46.028 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:46.029 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:46.062 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:46.161 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:46.161 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:46.280 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:46.281 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:46.323 INFO:teuthology.orchestra.run.vm05.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-08T23:39:46.323 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:46.323 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:46.323 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:46.324 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:46.325 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-08T23:39:46.325 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:46.347 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:46.347 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:46.380 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:46.447 INFO:teuthology.orchestra.run.vm02.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-08T23:39:46.447 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:46.447 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:46.447 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:46.447 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet zip 2026-03-08T23:39:46.448 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:46.471 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:46.471 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:46.505 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:46.603 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:46.603 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:46.721 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:46.722 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:46.818 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:46.818 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:46.819 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:46.820 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:46.820 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:46.820 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:46.820 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:46.820 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:46.820 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:46.820 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-08T23:39:46.820 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:46.835 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:46.835 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd* 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:46.910 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:46.911 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:46.911 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:46.911 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet zip 2026-03-08T23:39:46.911 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:46.921 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:46.921 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd* 2026-03-08T23:39:47.012 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-08T23:39:47.012 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 1186 kB disk space will be freed. 2026-03-08T23:39:47.050 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117410 files and directories currently installed.) 2026-03-08T23:39:47.052 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:47.117 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-08T23:39:47.117 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 1186 kB disk space will be freed. 2026-03-08T23:39:47.161 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117410 files and directories currently installed.) 2026-03-08T23:39:47.163 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:48.196 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:48.233 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:48.337 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:48.370 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:48.473 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:48.474 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:48.600 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:48.601 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:48.707 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:48.708 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:48.708 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:48.708 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-08T23:39:48.709 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:48.726 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:48.727 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-dev* libcephfs2* 2026-03-08T23:39:48.842 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:48.842 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:48.842 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:48.843 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:48.844 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:48.844 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet zip 2026-03-08T23:39:48.844 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:48.857 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:48.858 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-dev* libcephfs2* 2026-03-08T23:39:48.908 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-08T23:39:48.908 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 3202 kB disk space will be freed. 2026-03-08T23:39:48.950 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117402 files and directories currently installed.) 2026-03-08T23:39:48.952 INFO:teuthology.orchestra.run.vm05.stdout:Removing libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:48.965 INFO:teuthology.orchestra.run.vm05.stdout:Removing libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:48.993 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:39:49.069 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-08T23:39:49.069 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 3202 kB disk space will be freed. 2026-03-08T23:39:49.111 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117402 files and directories currently installed.) 2026-03-08T23:39:49.114 INFO:teuthology.orchestra.run.vm02.stdout:Removing libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:49.125 INFO:teuthology.orchestra.run.vm02.stdout:Removing libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:49.150 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:39:50.193 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:50.228 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:50.323 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:50.359 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:50.433 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:50.434 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:50.573 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:50.573 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:50.611 INFO:teuthology.orchestra.run.vm05.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-08T23:39:50.611 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:50.611 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:50.611 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:50.611 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:50.612 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:50.613 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:50.613 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet zip 2026-03-08T23:39:50.613 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:50.639 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:50.639 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:50.672 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:50.748 INFO:teuthology.orchestra.run.vm02.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-08T23:39:50.748 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:50.748 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:50.748 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-08T23:39:50.748 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet zip 2026-03-08T23:39:50.749 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:50.767 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:50.767 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:50.801 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:50.881 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:50.882 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:51.018 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:51.019 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:51.115 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:51.115 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:51.116 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-08T23:39:51.116 INFO:teuthology.orchestra.run.vm05.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:39:51.116 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-08T23:39:51.116 INFO:teuthology.orchestra.run.vm05.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-08T23:39:51.117 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:51.136 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:51.137 INFO:teuthology.orchestra.run.vm05.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-08T23:39:51.137 INFO:teuthology.orchestra.run.vm05.stdout: qemu-block-extra* rbd-fuse* 2026-03-08T23:39:51.216 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:51.217 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:51.218 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:51.218 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:51.218 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:51.218 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-08T23:39:51.218 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-08T23:39:51.218 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:51.233 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:51.233 INFO:teuthology.orchestra.run.vm02.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-08T23:39:51.234 INFO:teuthology.orchestra.run.vm02.stdout: qemu-block-extra* rbd-fuse* 2026-03-08T23:39:51.326 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-08T23:39:51.326 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 51.6 MB disk space will be freed. 2026-03-08T23:39:51.359 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117387 files and directories currently installed.) 2026-03-08T23:39:51.360 INFO:teuthology.orchestra.run.vm05.stdout:Removing rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.371 INFO:teuthology.orchestra.run.vm05.stdout:Removing libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.387 INFO:teuthology.orchestra.run.vm05.stdout:Removing libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.393 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-08T23:39:51.393 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 51.6 MB disk space will be freed. 2026-03-08T23:39:51.398 INFO:teuthology.orchestra.run.vm05.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-08T23:39:51.431 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117387 files and directories currently installed.) 2026-03-08T23:39:51.433 INFO:teuthology.orchestra.run.vm02.stdout:Removing rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.446 INFO:teuthology.orchestra.run.vm02.stdout:Removing libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.458 INFO:teuthology.orchestra.run.vm02.stdout:Removing libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.470 INFO:teuthology.orchestra.run.vm02.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-08T23:39:51.827 INFO:teuthology.orchestra.run.vm05.stdout:Removing librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.840 INFO:teuthology.orchestra.run.vm05.stdout:Removing librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.854 INFO:teuthology.orchestra.run.vm05.stdout:Removing librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.876 INFO:teuthology.orchestra.run.vm02.stdout:Removing librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.879 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:51.889 INFO:teuthology.orchestra.run.vm02.stdout:Removing librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.904 INFO:teuthology.orchestra.run.vm02.stdout:Removing librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:51.917 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:39:51.934 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:51.969 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:39:51.991 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-08T23:39:51.994 INFO:teuthology.orchestra.run.vm05.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-08T23:39:52.038 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-08T23:39:52.040 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-08T23:39:53.479 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:53.489 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:53.511 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:53.521 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:53.634 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:53.635 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:53.733 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:53.734 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:53.846 INFO:teuthology.orchestra.run.vm05.stdout:Package 'librbd1' is not installed, so not removed 2026-03-08T23:39:53.846 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:53.846 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:53.846 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-08T23:39:53.846 INFO:teuthology.orchestra.run.vm05.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:39:53.846 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-08T23:39:53.846 INFO:teuthology.orchestra.run.vm05.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-08T23:39:53.847 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:53.874 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:53.874 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:53.908 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:53.937 INFO:teuthology.orchestra.run.vm02.stdout:Package 'librbd1' is not installed, so not removed 2026-03-08T23:39:53.937 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:53.937 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:53.937 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-08T23:39:53.937 INFO:teuthology.orchestra.run.vm02.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:39:53.937 INFO:teuthology.orchestra.run.vm02.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:53.938 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:53.939 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:53.939 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-08T23:39:53.939 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-08T23:39:53.939 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:53.961 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:53.961 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:53.992 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:54.075 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:54.075 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:54.196 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:54.196 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:54.248 INFO:teuthology.orchestra.run.vm05.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-08T23:39:54.248 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:54.248 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:54.248 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-08T23:39:54.248 INFO:teuthology.orchestra.run.vm05.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:39:54.248 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-08T23:39:54.249 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:54.273 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:54.274 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:54.275 DEBUG:teuthology.orchestra.run.vm05:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-08T23:39:54.329 DEBUG:teuthology.orchestra.run.vm05:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:54.364 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:54.365 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:54.365 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:54.365 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:54.365 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:54.365 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:54.365 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-08T23:39:54.365 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-08T23:39:54.365 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-08T23:39:54.385 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-08T23:39:54.386 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:39:54.387 DEBUG:teuthology.orchestra.run.vm02:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-08T23:39:54.404 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:39:54.442 DEBUG:teuthology.orchestra.run.vm02:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-08T23:39:54.515 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:39:54.596 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-08T23:39:54.597 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-08T23:39:54.718 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-08T23:39:54.719 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-08T23:39:54.761 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be REMOVED: 2026-03-08T23:39:54.761 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:54.761 INFO:teuthology.orchestra.run.vm05.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-08T23:39:54.761 INFO:teuthology.orchestra.run.vm05.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:39:54.761 INFO:teuthology.orchestra.run.vm05.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-08T23:39:54.761 INFO:teuthology.orchestra.run.vm05.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-08T23:39:54.761 INFO:teuthology.orchestra.run.vm05.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-08T23:39:54.761 INFO:teuthology.orchestra.run.vm05.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-08T23:39:54.762 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-08T23:39:54.893 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-08T23:39:54.893 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-08T23:39:54.893 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-08T23:39:54.893 INFO:teuthology.orchestra.run.vm02.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-08T23:39:54.893 INFO:teuthology.orchestra.run.vm02.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-08T23:39:54.894 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-08T23:39:54.895 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-08T23:39:54.895 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-08T23:39:54.919 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 0 newly installed, 87 to remove and 10 not upgraded. 2026-03-08T23:39:54.919 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 107 MB disk space will be freed. 2026-03-08T23:39:54.957 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-08T23:39:54.960 INFO:teuthology.orchestra.run.vm05.stdout:Removing ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:54.976 INFO:teuthology.orchestra.run.vm05.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-08T23:39:54.988 INFO:teuthology.orchestra.run.vm05.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-08T23:39:55.000 INFO:teuthology.orchestra.run.vm05.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-08T23:39:55.011 INFO:teuthology.orchestra.run.vm05.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-08T23:39:55.022 INFO:teuthology.orchestra.run.vm05.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-08T23:39:55.032 INFO:teuthology.orchestra.run.vm05.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:39:55.044 INFO:teuthology.orchestra.run.vm05.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:39:55.054 INFO:teuthology.orchestra.run.vm05.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:39:55.069 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 87 to remove and 10 not upgraded. 2026-03-08T23:39:55.070 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 107 MB disk space will be freed. 2026-03-08T23:39:55.075 INFO:teuthology.orchestra.run.vm05.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-08T23:39:55.084 INFO:teuthology.orchestra.run.vm05.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-08T23:39:55.095 INFO:teuthology.orchestra.run.vm05.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-08T23:39:55.110 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-08T23:39:55.112 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:55.119 INFO:teuthology.orchestra.run.vm05.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-08T23:39:55.132 INFO:teuthology.orchestra.run.vm05.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-08T23:39:55.132 INFO:teuthology.orchestra.run.vm02.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-08T23:39:55.143 INFO:teuthology.orchestra.run.vm05.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-08T23:39:55.147 INFO:teuthology.orchestra.run.vm02.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-08T23:39:55.156 INFO:teuthology.orchestra.run.vm05.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-08T23:39:55.161 INFO:teuthology.orchestra.run.vm02.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-08T23:39:55.169 INFO:teuthology.orchestra.run.vm05.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-08T23:39:55.176 INFO:teuthology.orchestra.run.vm02.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-08T23:39:55.182 INFO:teuthology.orchestra.run.vm05.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-08T23:39:55.189 INFO:teuthology.orchestra.run.vm02.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-08T23:39:55.196 INFO:teuthology.orchestra.run.vm05.stdout:Removing luarocks (3.8.0+dfsg1-1) ... 2026-03-08T23:39:55.202 INFO:teuthology.orchestra.run.vm02.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:39:55.216 INFO:teuthology.orchestra.run.vm02.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:39:55.223 INFO:teuthology.orchestra.run.vm05.stdout:Removing liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-08T23:39:55.231 INFO:teuthology.orchestra.run.vm02.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-08T23:39:55.241 INFO:teuthology.orchestra.run.vm05.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-08T23:39:55.254 INFO:teuthology.orchestra.run.vm02.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-08T23:39:55.255 INFO:teuthology.orchestra.run.vm05.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-08T23:39:55.268 INFO:teuthology.orchestra.run.vm02.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-08T23:39:55.270 INFO:teuthology.orchestra.run.vm05.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-08T23:39:55.282 INFO:teuthology.orchestra.run.vm02.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-08T23:39:55.283 INFO:teuthology.orchestra.run.vm05.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-08T23:39:55.294 INFO:teuthology.orchestra.run.vm02.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-08T23:39:55.295 INFO:teuthology.orchestra.run.vm05.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-08T23:39:55.307 INFO:teuthology.orchestra.run.vm02.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-08T23:39:55.308 INFO:teuthology.orchestra.run.vm05.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-08T23:39:55.320 INFO:teuthology.orchestra.run.vm02.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-08T23:39:55.320 INFO:teuthology.orchestra.run.vm05.stdout:Removing libreadline-dev:amd64 (8.1.2-1) ... 2026-03-08T23:39:55.333 INFO:teuthology.orchestra.run.vm02.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-08T23:39:55.333 INFO:teuthology.orchestra.run.vm05.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-08T23:39:55.341 INFO:teuthology.orchestra.run.vm05.stdout:update-initramfs: deferring update (trigger activated) 2026-03-08T23:39:55.345 INFO:teuthology.orchestra.run.vm02.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-08T23:39:55.354 INFO:teuthology.orchestra.run.vm05.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-08T23:39:55.357 INFO:teuthology.orchestra.run.vm02.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-08T23:39:55.368 INFO:teuthology.orchestra.run.vm02.stdout:Removing luarocks (3.8.0+dfsg1-1) ... 2026-03-08T23:39:55.376 INFO:teuthology.orchestra.run.vm05.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-08T23:39:55.389 INFO:teuthology.orchestra.run.vm05.stdout:Removing lua-any (27ubuntu1) ... 2026-03-08T23:39:55.392 INFO:teuthology.orchestra.run.vm02.stdout:Removing liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-08T23:39:55.401 INFO:teuthology.orchestra.run.vm05.stdout:Removing lua-sec:amd64 (1.0.2-1) ... 2026-03-08T23:39:55.403 INFO:teuthology.orchestra.run.vm02.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-08T23:39:55.413 INFO:teuthology.orchestra.run.vm02.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-08T23:39:55.414 INFO:teuthology.orchestra.run.vm05.stdout:Removing lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-08T23:39:55.422 INFO:teuthology.orchestra.run.vm02.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-08T23:39:55.429 INFO:teuthology.orchestra.run.vm05.stdout:Removing lua5.1 (5.1.5-8.1build4) ... 2026-03-08T23:39:55.432 INFO:teuthology.orchestra.run.vm02.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-08T23:39:55.442 INFO:teuthology.orchestra.run.vm02.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-08T23:39:55.448 INFO:teuthology.orchestra.run.vm05.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-08T23:39:55.453 INFO:teuthology.orchestra.run.vm02.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-08T23:39:55.462 INFO:teuthology.orchestra.run.vm02.stdout:Removing libreadline-dev:amd64 (8.1.2-1) ... 2026-03-08T23:39:55.472 INFO:teuthology.orchestra.run.vm02.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-08T23:39:55.480 INFO:teuthology.orchestra.run.vm02.stdout:update-initramfs: deferring update (trigger activated) 2026-03-08T23:39:55.488 INFO:teuthology.orchestra.run.vm02.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-08T23:39:55.506 INFO:teuthology.orchestra.run.vm02.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-08T23:39:55.518 INFO:teuthology.orchestra.run.vm02.stdout:Removing lua-any (27ubuntu1) ... 2026-03-08T23:39:55.529 INFO:teuthology.orchestra.run.vm02.stdout:Removing lua-sec:amd64 (1.0.2-1) ... 2026-03-08T23:39:55.543 INFO:teuthology.orchestra.run.vm02.stdout:Removing lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-08T23:39:55.558 INFO:teuthology.orchestra.run.vm02.stdout:Removing lua5.1 (5.1.5-8.1build4) ... 2026-03-08T23:39:55.577 INFO:teuthology.orchestra.run.vm02.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-08T23:39:55.894 INFO:teuthology.orchestra.run.vm05.stdout:Removing pkg-config (0.29.2-1ubuntu3) ... 2026-03-08T23:39:55.928 INFO:teuthology.orchestra.run.vm05.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-08T23:39:55.956 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-pecan (1.3.3-4ubuntu2) ... 2026-03-08T23:39:55.968 INFO:teuthology.orchestra.run.vm02.stdout:Removing pkg-config (0.29.2-1ubuntu3) ... 2026-03-08T23:39:56.002 INFO:teuthology.orchestra.run.vm02.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-08T23:39:56.019 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-webtest (2.0.35-1) ... 2026-03-08T23:39:56.028 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-pecan (1.3.3-4ubuntu2) ... 2026-03-08T23:39:56.070 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-pastescript (2.0.2-4) ... 2026-03-08T23:39:56.093 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-webtest (2.0.35-1) ... 2026-03-08T23:39:56.124 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-pastedeploy (2.1.1-1) ... 2026-03-08T23:39:56.140 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-pastescript (2.0.2-4) ... 2026-03-08T23:39:56.174 INFO:teuthology.orchestra.run.vm05.stdout:Removing python-pastedeploy-tpl (2.1.1-1) ... 2026-03-08T23:39:56.186 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-08T23:39:56.192 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-pastedeploy (2.1.1-1) ... 2026-03-08T23:39:56.242 INFO:teuthology.orchestra.run.vm02.stdout:Removing python-pastedeploy-tpl (2.1.1-1) ... 2026-03-08T23:39:56.244 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-08T23:39:56.253 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-08T23:39:56.308 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-08T23:39:56.508 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-08T23:39:56.561 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-08T23:39:56.566 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-08T23:39:56.611 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:56.620 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-08T23:39:56.663 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:56.667 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:56.716 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-08T23:39:56.717 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-08T23:39:56.768 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-08T23:39:56.779 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-08T23:39:56.828 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-08T23:39:56.833 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-08T23:39:56.880 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-08T23:39:56.883 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-08T23:39:56.927 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-08T23:39:56.934 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-08T23:39:56.977 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-08T23:39:56.983 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-08T23:39:57.022 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-08T23:39:57.033 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-08T23:39:57.067 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-08T23:39:57.086 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-08T23:39:57.116 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-08T23:39:57.136 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-08T23:39:57.167 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-08T23:39:57.262 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-08T23:39:57.294 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-08T23:39:57.324 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-logutils (0.3.3-8) ... 2026-03-08T23:39:57.355 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-logutils (0.3.3-8) ... 2026-03-08T23:39:57.373 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-08T23:39:57.405 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-08T23:39:57.427 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-08T23:39:57.457 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-08T23:39:57.477 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-paste (3.5.0+dfsg1-1) ... 2026-03-08T23:39:57.511 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-paste (3.5.0+dfsg1-1) ... 2026-03-08T23:39:57.537 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-08T23:39:57.571 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-08T23:39:57.585 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-08T23:39:57.624 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-08T23:39:57.644 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-pyinotify (0.9.6-1.3) ... 2026-03-08T23:39:57.678 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-pyinotify (0.9.6-1.3) ... 2026-03-08T23:39:57.699 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-08T23:39:57.730 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-08T23:39:57.755 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-08T23:39:57.782 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-08T23:39:57.805 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-08T23:39:57.835 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-08T23:39:57.860 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-rsa (4.8-1) ... 2026-03-08T23:39:57.889 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-rsa (4.8-1) ... 2026-03-08T23:39:57.911 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-simplegeneric (0.8.1-3) ... 2026-03-08T23:39:57.942 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-simplegeneric (0.8.1-3) ... 2026-03-08T23:39:57.962 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-08T23:39:57.989 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-08T23:39:58.017 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-singledispatch (3.4.0.3-3) ... 2026-03-08T23:39:58.043 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-singledispatch (3.4.0.3-3) ... 2026-03-08T23:39:58.067 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-08T23:39:58.091 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-08T23:39:58.093 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-tempita (0.5.2-6ubuntu1) ... 2026-03-08T23:39:58.118 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-tempita (0.5.2-6ubuntu1) ... 2026-03-08T23:39:58.139 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-08T23:39:58.165 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-08T23:39:58.182 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-08T23:39:58.211 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-08T23:39:58.226 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-08T23:39:58.260 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-08T23:39:58.269 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-08T23:39:58.308 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-08T23:39:58.316 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-08T23:39:58.359 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-08T23:39:58.367 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-08T23:39:58.410 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-08T23:39:58.422 INFO:teuthology.orchestra.run.vm05.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-08T23:39:58.465 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-08T23:39:58.472 INFO:teuthology.orchestra.run.vm05.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-08T23:39:58.496 INFO:teuthology.orchestra.run.vm05.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-08T23:39:58.514 INFO:teuthology.orchestra.run.vm02.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-08T23:39:58.537 INFO:teuthology.orchestra.run.vm02.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-08T23:39:58.952 INFO:teuthology.orchestra.run.vm02.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-08T23:39:58.958 INFO:teuthology.orchestra.run.vm05.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-08T23:39:58.964 INFO:teuthology.orchestra.run.vm02.stdout:Removing unzip (6.0-26ubuntu3.2) ... 2026-03-08T23:39:58.971 INFO:teuthology.orchestra.run.vm05.stdout:Removing unzip (6.0-26ubuntu3.2) ... 2026-03-08T23:39:58.983 INFO:teuthology.orchestra.run.vm02.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-08T23:39:58.991 INFO:teuthology.orchestra.run.vm05.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-08T23:39:59.001 INFO:teuthology.orchestra.run.vm02.stdout:Removing zip (3.0-12build2) ... 2026-03-08T23:39:59.010 INFO:teuthology.orchestra.run.vm05.stdout:Removing zip (3.0-12build2) ... 2026-03-08T23:39:59.026 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:39:59.035 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:59.038 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-08T23:39:59.048 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-08T23:39:59.080 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-08T23:39:59.087 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-08T23:39:59.097 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-08T23:39:59.105 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-08T23:39:59.106 INFO:teuthology.orchestra.run.vm02.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-1092-kvm 2026-03-08T23:39:59.126 INFO:teuthology.orchestra.run.vm05.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-1092-kvm 2026-03-08T23:40:00.647 INFO:teuthology.orchestra.run.vm05.stdout:W: mkconf: MD subsystem is not loaded, thus I cannot scan for arrays. 2026-03-08T23:40:00.648 INFO:teuthology.orchestra.run.vm05.stdout:W: mdadm: failed to auto-generate temporary mdadm.conf file. 2026-03-08T23:40:00.670 INFO:teuthology.orchestra.run.vm02.stdout:W: mkconf: MD subsystem is not loaded, thus I cannot scan for arrays. 2026-03-08T23:40:00.672 INFO:teuthology.orchestra.run.vm02.stdout:W: mdadm: failed to auto-generate temporary mdadm.conf file. 2026-03-08T23:40:02.869 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:40:02.872 DEBUG:teuthology.parallel:result is None 2026-03-08T23:40:02.908 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-08T23:40:02.911 DEBUG:teuthology.parallel:result is None 2026-03-08T23:40:02.911 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm02.local 2026-03-08T23:40:02.911 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm05.local 2026-03-08T23:40:02.911 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-08T23:40:02.911 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-08T23:40:02.922 DEBUG:teuthology.orchestra.run.vm02:> sudo apt-get update 2026-03-08T23:40:02.960 DEBUG:teuthology.orchestra.run.vm05:> sudo apt-get update 2026-03-08T23:40:03.192 INFO:teuthology.orchestra.run.vm05.stdout:Hit:1 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-08T23:40:03.193 INFO:teuthology.orchestra.run.vm02.stdout:Hit:1 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-08T23:40:03.196 INFO:teuthology.orchestra.run.vm05.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-08T23:40:03.197 INFO:teuthology.orchestra.run.vm02.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-08T23:40:03.204 INFO:teuthology.orchestra.run.vm05.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-08T23:40:03.205 INFO:teuthology.orchestra.run.vm02.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-08T23:40:03.259 INFO:teuthology.orchestra.run.vm05.stdout:Hit:4 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-08T23:40:03.506 INFO:teuthology.orchestra.run.vm02.stdout:Hit:4 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-08T23:40:04.160 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-08T23:40:04.172 DEBUG:teuthology.parallel:result is None 2026-03-08T23:40:04.357 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-08T23:40:04.370 DEBUG:teuthology.parallel:result is None 2026-03-08T23:40:04.370 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-08T23:40:04.373 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-08T23:40:04.373 DEBUG:teuthology.orchestra.run.vm02:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-08T23:40:04.374 DEBUG:teuthology.orchestra.run.vm05:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout: remote refid st t when poll reach delay offset jitter 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================== 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+mail.vbrandl.ne 71.58.123.92 3 u 26 64 177 25.046 -0.186 2.980 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+de.relay.mahi.b 232.208.203.34 3 u 24 64 177 21.054 +6.888 6.357 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+185.13.148.71 79.133.44.146 2 u 29 64 177 32.137 +0.813 2.541 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+vps-fra2.orlean 169.254.169.254 4 u 27 64 177 21.011 +0.713 2.576 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+de.timeserver.p 131.188.3.221 2 u 26 64 177 25.217 +0.738 2.414 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+152.53.184.199 131.188.3.220 2 u 30 64 177 32.921 -1.346 3.008 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+obelix.hetzner. 213.239.239.166 3 u 25 64 177 25.089 -1.030 2.623 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+77.90.16.198 46.224.156.215 3 u 25 64 177 24.508 +5.594 5.019 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+v22025082392863 129.69.253.1 2 u 28 64 177 28.266 -1.936 2.536 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:*141.144.246.224 146.131.121.246 2 u 26 64 177 29.472 +0.534 2.079 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+ns.gunnarhofman 237.17.204.95 2 u 25 64 177 24.993 +3.651 4.057 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout: mail3.kekew.inf 20.101.57.9 4 u 31 64 177 41.282 +2.107 3.513 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+alphyn.canonica 132.163.96.1 2 u 46 64 177 103.509 -1.502 3.901 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+139-144-71-56.i 82.35.162.146 2 u 29 64 177 22.731 -3.447 2.286 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+mx10.edv-hueske 189.97.54.122 2 u 28 64 177 28.816 +0.573 2.118 2026-03-08T23:40:06.580 INFO:teuthology.orchestra.run.vm02.stdout:+185.125.190.57 194.121.207.249 2 u 39 64 177 31.212 +1.307 2.636 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout: remote refid st t when poll reach delay offset jitter 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:============================================================================== 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:#mail.vbrandl.ne 71.58.123.92 3 u 32 64 177 25.055 -0.919 0.968 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:-de.relay.mahi.b 232.208.203.34 3 u 30 64 177 21.186 +0.699 1.006 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:-v22025082392863 129.69.253.1 2 u 30 64 177 28.615 -2.838 1.254 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:+obelix.hetzner. 213.239.239.166 3 u 28 64 177 25.221 -0.060 1.131 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:+141.144.246.224 146.131.121.246 2 u 32 64 177 29.437 -0.383 1.195 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:+185.13.148.71 79.133.44.146 2 u 34 64 177 32.062 +0.035 0.983 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:#ns.gunnarhofman 237.17.204.95 2 u 29 64 177 24.923 -0.123 1.146 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:-de.timeserver.p 131.188.3.221 2 u 29 64 177 26.399 -0.610 1.407 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:-m1000-ntp-pool. .MRS. 1 u 26 64 177 28.884 +1.759 0.452 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:-vps-fra2.orlean 169.254.169.254 4 u 25 64 177 20.958 +0.978 0.411 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:*mx10.edv-hueske 189.97.54.122 2 u 26 64 177 28.755 -0.193 1.119 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:#alphyn.canonica 132.163.96.1 2 u 38 64 177 102.166 -2.637 1.077 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:#139-144-71-56.i 82.35.162.146 2 u 27 64 177 22.701 -4.449 0.745 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:-77.90.16.198 46.224.156.215 3 u 21 64 177 24.513 +1.834 0.419 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:#185.125.190.56 79.243.60.50 2 u 42 64 177 36.962 -0.677 1.067 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout:#185.125.190.58 145.238.80.80 2 u 37 64 177 35.461 -1.878 1.272 2026-03-08T23:40:10.559 INFO:teuthology.orchestra.run.vm05.stdout: mail3.kekew.inf 20.101.57.9 4 u 28 64 177 39.742 +2.039 0.528 2026-03-08T23:40:10.560 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-08T23:40:10.562 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-08T23:40:10.562 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-08T23:40:10.565 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-08T23:40:10.568 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-08T23:40:10.571 INFO:teuthology.task.internal:Duration was 577.144334 seconds 2026-03-08T23:40:10.572 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-08T23:40:10.574 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-08T23:40:10.574 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-08T23:40:10.576 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-08T23:40:10.603 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-08T23:40:10.603 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm02.local 2026-03-08T23:40:10.603 DEBUG:teuthology.orchestra.run.vm02:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-08T23:40:10.654 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm05.local 2026-03-08T23:40:10.654 DEBUG:teuthology.orchestra.run.vm05:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-08T23:40:10.666 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-08T23:40:10.666 DEBUG:teuthology.orchestra.run.vm02:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-08T23:40:10.695 DEBUG:teuthology.orchestra.run.vm05:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-08T23:40:10.745 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-08T23:40:10.745 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-08T23:40:10.775 DEBUG:teuthology.orchestra.run.vm05:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-08T23:40:10.782 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-08T23:40:10.782 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-08T23:40:10.783 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5/home/ubuntu/cephtest/archive/syslog/misc.log: --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-08T23:40:10.783 INFO:teuthology.orchestra.run.vm02.stderr: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-08T23:40:10.783 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-08T23:40:10.792 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 88.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-08T23:40:10.795 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-08T23:40:10.796 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-08T23:40:10.796 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0%gzip -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-08T23:40:10.796 INFO:teuthology.orchestra.run.vm05.stderr: -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-08T23:40:10.796 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-08T23:40:10.804 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 88.1% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-08T23:40:10.805 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-08T23:40:10.807 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-08T23:40:10.808 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-08T23:40:10.841 DEBUG:teuthology.orchestra.run.vm05:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-08T23:40:10.857 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-08T23:40:10.859 DEBUG:teuthology.orchestra.run.vm02:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-08T23:40:10.883 DEBUG:teuthology.orchestra.run.vm05:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-08T23:40:10.889 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = core 2026-03-08T23:40:10.907 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = core 2026-03-08T23:40:10.914 DEBUG:teuthology.orchestra.run.vm02:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-08T23:40:10.940 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:40:10.941 DEBUG:teuthology.orchestra.run.vm05:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-08T23:40:10.959 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-08T23:40:10.959 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-08T23:40:10.963 INFO:teuthology.task.internal:Transferring archived files... 2026-03-08T23:40:10.963 DEBUG:teuthology.misc:Transferring archived files from vm02:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300/remote/vm02 2026-03-08T23:40:10.963 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-08T23:40:10.990 DEBUG:teuthology.misc:Transferring archived files from vm05:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-08_22:22:45-orch:cephadm-squid-none-default-vps/300/remote/vm05 2026-03-08T23:40:10.991 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-08T23:40:11.008 INFO:teuthology.task.internal:Removing archive directory... 2026-03-08T23:40:11.008 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-08T23:40:11.035 DEBUG:teuthology.orchestra.run.vm05:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-08T23:40:11.055 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-08T23:40:11.058 INFO:teuthology.task.internal:Not uploading archives. 2026-03-08T23:40:11.058 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-08T23:40:11.061 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-08T23:40:11.061 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-08T23:40:11.079 DEBUG:teuthology.orchestra.run.vm05:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-08T23:40:11.081 INFO:teuthology.orchestra.run.vm02.stdout: 258077 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 8 23:40 /home/ubuntu/cephtest 2026-03-08T23:40:11.099 INFO:teuthology.orchestra.run.vm05.stdout: 258069 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 8 23:40 /home/ubuntu/cephtest 2026-03-08T23:40:11.100 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-08T23:40:11.106 INFO:teuthology.run:Summary data: description: orch:cephadm/workunits/{0-distro/ubuntu_22.04 agent/off mon_election/classic task/test_extra_daemon_features} duration: 577.1443336009979 flavor: default owner: kyr success: true 2026-03-08T23:40:11.106 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-08T23:40:11.132 INFO:teuthology.run:pass