2026-03-31T11:28:50.634 INFO:root:teuthology version: 1.2.4.dev37+ga59626679 2026-03-31T11:28:50.639 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-31T11:28:50.660 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300 branch: tentacle description: rados/cephadm/workunits/{0-distro/ubuntu_22.04 agent/on mon_election/connectivity task/test_extra_daemon_features} email: null first_in_suite: false flavor: default job_id: '4300' ktype: distro last_in_suite: false machine_type: vps name: kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps no_nested_subset: false os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: tentacle ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: logical_volumes: lv_1: scratch_dev: true size: 25%VG vg: vg_nvme lv_2: scratch_dev: true size: 25%VG vg: vg_nvme lv_3: scratch_dev: true size: 25%VG vg: vg_nvme lv_4: scratch_dev: true size: 25%VG vg: vg_nvme timezone: UTC volume_groups: vg_nvme: pvs: /dev/vdb,/dev/vdc,/dev/vdd,/dev/vde ceph: conf: global: mon election default strategy: 3 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: true mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: 5bb3278730741031382ca9c3dc9d221a942e06a2 ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm install: ceph: flavor: default sha1: 5bb3278730741031382ca9c3dc9d221a942e06a2 extra_system_packages: deb: - python3-jmespath - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-jmespath - python3-xmltodict - s3cmd workunit: branch: tt-tentacle sha1: 0392f78529848ec72469e8e431875cb98d3a5fb4 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 seed: 6407 sha1: 5bb3278730741031382ca9c3dc9d221a942e06a2 sleep_before_teardown: 0 subset: 1/100000 suite: rados suite_branch: tt-tentacle suite_path: /home/teuthos/src/github.com_kshtsk_ceph_0392f78529848ec72469e8e431875cb98d3a5fb4/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 0392f78529848ec72469e8e431875cb98d3a5fb4 targets: vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCA2L9mMTbUCl4kO2VJbdeSh5m43Q0B6ltPN8Li9buOpOYaXIkcvjE85gS1cRBFnMTiMp54Q/WQXSoxe20Fvy+I= vm09.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKKIGUi8dIAxu9WQ5ZhhVYHM3ZoghSZtEu9E3ue6BPeK5bKVaZDszNmTqeYWQfYH78f5iMNhRc1Xz8J5MeVV0IQ= tasks: - install: null - cephadm: null - exec: all-hosts: - mkdir /etc/cephadm_testing - cephadm.apply: specs: - extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon - custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest - custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z - cephadm.wait_for_service: service: mon - cephadm.wait_for_service: service: container.foo - cephadm.wait_for_service: service: container.bar - exec: host.a: - 'set -ex FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) sleep 60 # check extra container and entrypoint args written to mon unit run file grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run # check that custom container properly wrote content to file. # This requires the custom config, extra container args, and # entrypoint args to all be working in order for this to have # been written. The container entrypoint was set up with custom_configs, # the content and where to write to with the entrypoint args, and the mounting # of the /etc/cephadm_testing dir with extra container args grep "testing_custom_containers" /etc/cephadm_testing/testing.txt # Verify that container bar''s init containers and primary container # ran successfully dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name ''container.bar.*'') test -n "$dir" grep ok ${dir}/data/primary.txt grep from=centos ${dir}/data/from.txt test -s ${dir}/data/presized.dat ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: uv2 teuthology_repo: https://github.com/kshtsk/teuthology teuthology_sha1: a59626679648f962bca99d20d35578f2998c8f37 timestamp: 2026-03-31_11:18:10 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.282426 2026-03-31T11:28:50.660 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_0392f78529848ec72469e8e431875cb98d3a5fb4/qa; will attempt to use it 2026-03-31T11:28:50.660 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_0392f78529848ec72469e8e431875cb98d3a5fb4/qa/tasks 2026-03-31T11:28:50.660 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-31T11:28:50.660 INFO:teuthology.task.internal:Checking packages... 2026-03-31T11:28:50.660 INFO:teuthology.task.internal:Checking packages for os_type 'ubuntu', flavor 'default' and ceph hash '5bb3278730741031382ca9c3dc9d221a942e06a2' 2026-03-31T11:28:50.660 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-31T11:28:50.660 INFO:teuthology.packaging:ref: None 2026-03-31T11:28:50.660 INFO:teuthology.packaging:tag: None 2026-03-31T11:28:50.660 INFO:teuthology.packaging:branch: tentacle 2026-03-31T11:28:50.660 INFO:teuthology.packaging:sha1: 5bb3278730741031382ca9c3dc9d221a942e06a2 2026-03-31T11:28:50.660 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&ref=tentacle 2026-03-31T11:28:51.405 INFO:teuthology.task.internal:Found packages for ceph version 20.2.0-714-g147f7c6a-1jammy 2026-03-31T11:28:51.405 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-31T11:28:51.405 INFO:teuthology.task.internal:no buildpackages task found 2026-03-31T11:28:51.405 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-31T11:28:51.406 INFO:teuthology.task.internal:Saving configuration 2026-03-31T11:28:51.412 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-31T11:28:51.413 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-31T11:28:51.419 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm06.local', 'description': '/archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-31 11:27:47.320316', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:06', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCA2L9mMTbUCl4kO2VJbdeSh5m43Q0B6ltPN8Li9buOpOYaXIkcvjE85gS1cRBFnMTiMp54Q/WQXSoxe20Fvy+I='} 2026-03-31T11:28:51.423 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm09.local', 'description': '/archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-31 11:27:47.320735', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:09', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKKIGUi8dIAxu9WQ5ZhhVYHM3ZoghSZtEu9E3ue6BPeK5bKVaZDszNmTqeYWQfYH78f5iMNhRc1Xz8J5MeVV0IQ='} 2026-03-31T11:28:51.423 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-31T11:28:51.423 INFO:teuthology.task.internal:roles: ubuntu@vm06.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0'] 2026-03-31T11:28:51.423 INFO:teuthology.task.internal:roles: ubuntu@vm09.local - ['host.b', 'mon.b', 'mgr.b', 'osd.1'] 2026-03-31T11:28:51.423 INFO:teuthology.run_tasks:Running task console_log... 2026-03-31T11:28:51.429 DEBUG:teuthology.task.console_log:vm06 does not support IPMI; excluding 2026-03-31T11:28:51.432 DEBUG:teuthology.task.console_log:vm09 does not support IPMI; excluding 2026-03-31T11:28:51.433 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f171664dbd0>, signals=[15]) 2026-03-31T11:28:51.433 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-31T11:28:51.433 INFO:teuthology.task.internal:Opening connections... 2026-03-31T11:28:51.433 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-03-31T11:28:51.448 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-31T11:28:51.508 DEBUG:teuthology.task.internal:connecting to ubuntu@vm09.local 2026-03-31T11:28:51.508 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-31T11:28:51.569 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-31T11:28:51.570 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-03-31T11:28:51.602 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-03-31T11:28:51.602 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-03-31T11:28:51.645 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-31T11:28:51.645 INFO:teuthology.orchestra.run.vm06.stdout:NAME="Ubuntu" 2026-03-31T11:28:51.645 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="22.04" 2026-03-31T11:28:51.645 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-31T11:28:51.645 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_CODENAME=jammy 2026-03-31T11:28:51.645 INFO:teuthology.orchestra.run.vm06.stdout:ID=ubuntu 2026-03-31T11:28:51.645 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE=debian 2026-03-31T11:28:51.645 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-31T11:28:51.646 INFO:teuthology.orchestra.run.vm06.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-31T11:28:51.646 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-31T11:28:51.646 INFO:teuthology.orchestra.run.vm06.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-31T11:28:51.646 INFO:teuthology.orchestra.run.vm06.stdout:UBUNTU_CODENAME=jammy 2026-03-31T11:28:51.646 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-03-31T11:28:51.650 DEBUG:teuthology.orchestra.run.vm09:> uname -m 2026-03-31T11:28:51.653 INFO:teuthology.orchestra.run.vm09.stdout:x86_64 2026-03-31T11:28:51.653 DEBUG:teuthology.orchestra.run.vm09:> cat /etc/os-release 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:NAME="Ubuntu" 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_ID="22.04" 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_CODENAME=jammy 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:ID=ubuntu 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:ID_LIKE=debian 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-31T11:28:51.699 INFO:teuthology.orchestra.run.vm09.stdout:UBUNTU_CODENAME=jammy 2026-03-31T11:28:51.699 INFO:teuthology.lock.ops:Updating vm09.local on lock server 2026-03-31T11:28:51.703 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-31T11:28:51.705 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-31T11:28:51.706 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-31T11:28:51.706 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-03-31T11:28:51.707 DEBUG:teuthology.orchestra.run.vm09:> test '!' -e /home/ubuntu/cephtest 2026-03-31T11:28:51.742 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-31T11:28:51.743 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-31T11:28:51.743 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-03-31T11:28:51.751 DEBUG:teuthology.orchestra.run.vm09:> test -z $(ls -A /var/lib/ceph) 2026-03-31T11:28:51.753 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-31T11:28:51.786 INFO:teuthology.orchestra.run.vm09.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-31T11:28:51.787 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-31T11:28:51.794 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-03-31T11:28:51.797 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-31T11:28:52.029 DEBUG:teuthology.orchestra.run.vm09:> test -e /ceph-qa-ready 2026-03-31T11:28:52.032 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-31T11:28:52.256 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-31T11:28:52.257 INFO:teuthology.task.internal:Creating test directory... 2026-03-31T11:28:52.258 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-31T11:28:52.259 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-31T11:28:52.261 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-31T11:28:52.262 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-31T11:28:52.263 INFO:teuthology.task.internal:Creating archive directory... 2026-03-31T11:28:52.264 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-31T11:28:52.303 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-31T11:28:52.312 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-31T11:28:52.313 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-31T11:28:52.313 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-31T11:28:52.349 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-31T11:28:52.349 DEBUG:teuthology.orchestra.run.vm09:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-31T11:28:52.354 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-31T11:28:52.354 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-31T11:28:52.391 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-31T11:28:52.399 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-31T11:28:52.403 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-31T11:28:52.403 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-31T11:28:52.407 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-31T11:28:52.408 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-31T11:28:52.410 INFO:teuthology.task.internal:Configuring sudo... 2026-03-31T11:28:52.410 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-31T11:28:52.447 DEBUG:teuthology.orchestra.run.vm09:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-31T11:28:52.459 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-31T11:28:52.461 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-31T11:28:52.461 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-31T11:28:52.496 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-31T11:28:52.503 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-31T11:28:52.541 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-31T11:28:52.585 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:28:52.585 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-31T11:28:52.634 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-31T11:28:52.637 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-31T11:28:52.683 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:28:52.683 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-31T11:28:52.731 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-03-31T11:28:52.732 DEBUG:teuthology.orchestra.run.vm09:> sudo service rsyslog restart 2026-03-31T11:28:52.786 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-31T11:28:52.788 INFO:teuthology.task.internal:Starting timer... 2026-03-31T11:28:52.788 INFO:teuthology.run_tasks:Running task pcp... 2026-03-31T11:28:52.790 INFO:teuthology.run_tasks:Running task selinux... 2026-03-31T11:28:52.792 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-03-31T11:28:52.793 INFO:teuthology.task.selinux:Excluding vm09: VMs are not yet supported 2026-03-31T11:28:52.793 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-31T11:28:52.793 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-31T11:28:52.793 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-31T11:28:52.793 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-31T11:28:52.794 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}} 2026-03-31T11:28:52.794 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-31T11:28:52.795 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-31T11:28:53.496 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-31T11:28:53.501 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-31T11:28:53.501 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "logical_volumes": {"lv_1": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_2": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_3": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_4": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}}, "timezone": "UTC", "volume_groups": {"vg_nvme": {"pvs": "/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde"}}}' -i /tmp/teuth_ansible_inventory3sx4ydzu --limit vm06.local,vm09.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-31T11:31:03.467 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm06.local'), Remote(name='ubuntu@vm09.local')] 2026-03-31T11:31:03.467 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-03-31T11:31:03.468 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-31T11:31:03.528 DEBUG:teuthology.orchestra.run.vm06:> true 2026-03-31T11:31:03.745 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-03-31T11:31:03.745 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm09.local' 2026-03-31T11:31:03.745 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-31T11:31:03.805 DEBUG:teuthology.orchestra.run.vm09:> true 2026-03-31T11:31:04.020 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm09.local' 2026-03-31T11:31:04.021 INFO:teuthology.run_tasks:Running task clock... 2026-03-31T11:31:04.023 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-31T11:31:04.023 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-31T11:31:04.023 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-31T11:31:04.024 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-31T11:31:04.024 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-31T11:31:04.040 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-31T11:31:04.040 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Command line: ntpd -gq 2026-03-31T11:31:04.040 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: ---------------------------------------------------- 2026-03-31T11:31:04.040 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: ntp-4 is maintained by Network Time Foundation, 2026-03-31T11:31:04.040 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-31T11:31:04.040 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: corporation. Support and training for ntp-4 are 2026-03-31T11:31:04.040 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: available at https://www.nwtime.org/support 2026-03-31T11:31:04.040 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: ---------------------------------------------------- 2026-03-31T11:31:04.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: proto: precision = 0.029 usec (-25) 2026-03-31T11:31:04.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: basedate set to 2022-02-04 2026-03-31T11:31:04.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: gps base set to 2022-02-06 (week 2196) 2026-03-31T11:31:04.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-31T11:31:04.042 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-31T11:31:04.042 INFO:teuthology.orchestra.run.vm06.stderr:31 Mar 11:31:04 ntpd[16229]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 94 days ago 2026-03-31T11:31:04.042 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Listen and drop on 0 v6wildcard [::]:123 2026-03-31T11:31:04.043 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-31T11:31:04.043 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Listen normally on 2 lo 127.0.0.1:123 2026-03-31T11:31:04.043 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Listen normally on 3 ens3 192.168.123.106:123 2026-03-31T11:31:04.043 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Listen normally on 4 lo [::1]:123 2026-03-31T11:31:04.043 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:6%2]:123 2026-03-31T11:31:04.043 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:04 ntpd[16229]: Listening on routing socket on fd #22 for interface updates 2026-03-31T11:31:04.078 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Command line: ntpd -gq 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: ---------------------------------------------------- 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: ntp-4 is maintained by Network Time Foundation, 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: corporation. Support and training for ntp-4 are 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: available at https://www.nwtime.org/support 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: ---------------------------------------------------- 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: proto: precision = 0.029 usec (-25) 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: basedate set to 2022-02-04 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: gps base set to 2022-02-06 (week 2196) 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-31T11:31:04.079 INFO:teuthology.orchestra.run.vm09.stderr:31 Mar 11:31:04 ntpd[16225]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 94 days ago 2026-03-31T11:31:04.080 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Listen and drop on 0 v6wildcard [::]:123 2026-03-31T11:31:04.080 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-31T11:31:04.080 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Listen normally on 2 lo 127.0.0.1:123 2026-03-31T11:31:04.080 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Listen normally on 3 ens3 192.168.123.109:123 2026-03-31T11:31:04.080 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Listen normally on 4 lo [::1]:123 2026-03-31T11:31:04.080 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:9%2]:123 2026-03-31T11:31:04.080 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:04 ntpd[16225]: Listening on routing socket on fd #22 for interface updates 2026-03-31T11:31:05.042 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:05 ntpd[16229]: Soliciting pool server 31.209.85.243 2026-03-31T11:31:05.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:05 ntpd[16225]: Soliciting pool server 31.209.85.243 2026-03-31T11:31:06.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:06 ntpd[16229]: Soliciting pool server 217.14.146.53 2026-03-31T11:31:06.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:06 ntpd[16229]: Soliciting pool server 176.9.8.206 2026-03-31T11:31:06.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:06 ntpd[16225]: Soliciting pool server 217.14.146.53 2026-03-31T11:31:06.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:06 ntpd[16225]: Soliciting pool server 176.9.8.206 2026-03-31T11:31:07.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:07 ntpd[16229]: Soliciting pool server 90.187.112.137 2026-03-31T11:31:07.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:07 ntpd[16229]: Soliciting pool server 185.16.60.96 2026-03-31T11:31:07.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:07 ntpd[16225]: Soliciting pool server 90.187.112.137 2026-03-31T11:31:07.191 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:07 ntpd[16229]: Soliciting pool server 93.177.65.20 2026-03-31T11:31:07.191 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:07 ntpd[16225]: Soliciting pool server 93.177.65.20 2026-03-31T11:31:08.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:08 ntpd[16229]: Soliciting pool server 131.188.3.223 2026-03-31T11:31:08.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:08 ntpd[16229]: Soliciting pool server 168.119.211.223 2026-03-31T11:31:08.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:08 ntpd[16229]: Soliciting pool server 185.252.140.126 2026-03-31T11:31:08.042 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:08 ntpd[16229]: Soliciting pool server 77.42.16.222 2026-03-31T11:31:08.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:08 ntpd[16225]: Soliciting pool server 131.188.3.223 2026-03-31T11:31:08.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:08 ntpd[16225]: Soliciting pool server 168.119.211.223 2026-03-31T11:31:08.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:08 ntpd[16225]: Soliciting pool server 77.42.16.222 2026-03-31T11:31:09.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:09 ntpd[16229]: Soliciting pool server 128.140.109.119 2026-03-31T11:31:09.042 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:09 ntpd[16229]: Soliciting pool server 93.241.86.156 2026-03-31T11:31:09.042 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:09 ntpd[16229]: Soliciting pool server 85.10.240.253 2026-03-31T11:31:09.042 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:09 ntpd[16229]: Soliciting pool server 91.189.91.157 2026-03-31T11:31:09.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:09 ntpd[16225]: Soliciting pool server 128.140.109.119 2026-03-31T11:31:09.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:09 ntpd[16225]: Soliciting pool server 93.241.86.156 2026-03-31T11:31:09.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:09 ntpd[16225]: Soliciting pool server 85.10.240.253 2026-03-31T11:31:09.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:09 ntpd[16225]: Soliciting pool server 91.189.91.157 2026-03-31T11:31:10.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:10 ntpd[16229]: Soliciting pool server 185.125.190.58 2026-03-31T11:31:10.041 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:10 ntpd[16229]: Soliciting pool server 78.47.56.71 2026-03-31T11:31:10.042 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:10 ntpd[16229]: Soliciting pool server 85.215.122.93 2026-03-31T11:31:10.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:10 ntpd[16225]: Soliciting pool server 185.125.190.58 2026-03-31T11:31:10.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:10 ntpd[16225]: Soliciting pool server 78.47.56.71 2026-03-31T11:31:10.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:10 ntpd[16225]: Soliciting pool server 85.215.122.93 2026-03-31T11:31:11.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:11 ntpd[16225]: Soliciting pool server 185.125.190.57 2026-03-31T11:31:11.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:11 ntpd[16225]: Soliciting pool server 148.251.54.81 2026-03-31T11:31:11.079 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:11 ntpd[16225]: Soliciting pool server 2003:a:42b:e400::3 2026-03-31T11:31:13.072 INFO:teuthology.orchestra.run.vm06.stdout:31 Mar 11:31:13 ntpd[16229]: ntpd: time slew +0.005975 s 2026-03-31T11:31:13.072 INFO:teuthology.orchestra.run.vm06.stdout:ntpd: time slew +0.005975s 2026-03-31T11:31:13.093 INFO:teuthology.orchestra.run.vm06.stdout: remote refid st t when poll reach delay offset jitter 2026-03-31T11:31:13.093 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================== 2026-03-31T11:31:13.093 INFO:teuthology.orchestra.run.vm06.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.093 INFO:teuthology.orchestra.run.vm06.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.093 INFO:teuthology.orchestra.run.vm06.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.093 INFO:teuthology.orchestra.run.vm06.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.093 INFO:teuthology.orchestra.run.vm06.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.110 INFO:teuthology.orchestra.run.vm09.stdout:31 Mar 11:31:13 ntpd[16225]: ntpd: time slew +0.005873 s 2026-03-31T11:31:13.110 INFO:teuthology.orchestra.run.vm09.stdout:ntpd: time slew +0.005873s 2026-03-31T11:31:13.131 INFO:teuthology.orchestra.run.vm09.stdout: remote refid st t when poll reach delay offset jitter 2026-03-31T11:31:13.131 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================== 2026-03-31T11:31:13.131 INFO:teuthology.orchestra.run.vm09.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.131 INFO:teuthology.orchestra.run.vm09.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.131 INFO:teuthology.orchestra.run.vm09.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.131 INFO:teuthology.orchestra.run.vm09.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.131 INFO:teuthology.orchestra.run.vm09.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:31:13.131 INFO:teuthology.run_tasks:Running task install... 2026-03-31T11:31:13.133 DEBUG:teuthology.task.install:project ceph 2026-03-31T11:31:13.133 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': '5bb3278730741031382ca9c3dc9d221a942e06a2'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}} 2026-03-31T11:31:13.133 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': '5bb3278730741031382ca9c3dc9d221a942e06a2', 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}} 2026-03-31T11:31:13.133 INFO:teuthology.task.install:Using flavor: default 2026-03-31T11:31:13.136 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-31T11:31:13.136 INFO:teuthology.task.install:extra packages: [] 2026-03-31T11:31:13.136 DEBUG:teuthology.orchestra.run.vm06:> sudo apt-key list | grep Ceph 2026-03-31T11:31:13.136 DEBUG:teuthology.orchestra.run.vm09:> sudo apt-key list | grep Ceph 2026-03-31T11:31:13.177 INFO:teuthology.orchestra.run.vm06.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-31T11:31:13.197 INFO:teuthology.orchestra.run.vm06.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-31T11:31:13.197 INFO:teuthology.orchestra.run.vm06.stdout:uid [ unknown] Ceph.com (release key) 2026-03-31T11:31:13.198 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-31T11:31:13.198 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-31T11:31:13.198 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=5bb3278730741031382ca9c3dc9d221a942e06a2 2026-03-31T11:31:13.211 INFO:teuthology.orchestra.run.vm09.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-31T11:31:13.231 INFO:teuthology.orchestra.run.vm09.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-31T11:31:13.231 INFO:teuthology.orchestra.run.vm09.stdout:uid [ unknown] Ceph.com (release key) 2026-03-31T11:31:13.231 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-31T11:31:13.231 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-31T11:31:13.231 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=5bb3278730741031382ca9c3dc9d221a942e06a2 2026-03-31T11:31:13.749 INFO:teuthology.task.install.deb:Pulling from https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default/ 2026-03-31T11:31:13.749 INFO:teuthology.task.install.deb:Package version is 20.2.0-721-g5bb32787-1jammy 2026-03-31T11:31:13.812 INFO:teuthology.task.install.deb:Pulling from https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default/ 2026-03-31T11:31:13.812 INFO:teuthology.task.install.deb:Package version is 20.2.0-721-g5bb32787-1jammy 2026-03-31T11:31:14.291 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:31:14.291 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-31T11:31:14.299 DEBUG:teuthology.orchestra.run.vm06:> sudo apt-get update 2026-03-31T11:31:14.343 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:31:14.343 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-31T11:31:14.353 DEBUG:teuthology.orchestra.run.vm09:> sudo apt-get update 2026-03-31T11:31:14.607 INFO:teuthology.orchestra.run.vm06.stdout:Hit:1 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-31T11:31:14.607 INFO:teuthology.orchestra.run.vm06.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-31T11:31:14.676 INFO:teuthology.orchestra.run.vm09.stdout:Hit:1 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-31T11:31:14.676 INFO:teuthology.orchestra.run.vm09.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-31T11:31:14.706 INFO:teuthology.orchestra.run.vm06.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-31T11:31:14.786 INFO:teuthology.orchestra.run.vm09.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-31T11:31:14.810 INFO:teuthology.orchestra.run.vm06.stdout:Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-31T11:31:14.896 INFO:teuthology.orchestra.run.vm09.stdout:Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-31T11:31:14.971 INFO:teuthology.orchestra.run.vm06.stdout:Ign:5 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy InRelease 2026-03-31T11:31:14.987 INFO:teuthology.orchestra.run.vm09.stdout:Ign:5 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy InRelease 2026-03-31T11:31:15.091 INFO:teuthology.orchestra.run.vm06.stdout:Get:6 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy Release [7689 B] 2026-03-31T11:31:15.102 INFO:teuthology.orchestra.run.vm09.stdout:Get:6 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy Release [7689 B] 2026-03-31T11:31:15.200 INFO:teuthology.orchestra.run.vm06.stdout:Ign:7 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-31T11:31:15.218 INFO:teuthology.orchestra.run.vm09.stdout:Ign:7 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-31T11:31:15.395 INFO:teuthology.orchestra.run.vm06.stdout:Get:8 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-31T11:31:15.399 INFO:teuthology.orchestra.run.vm09.stdout:Get:8 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-31T11:31:15.474 INFO:teuthology.orchestra.run.vm06.stdout:Fetched 26.5 kB in 1s (26.1 kB/s) 2026-03-31T11:31:15.476 INFO:teuthology.orchestra.run.vm09.stdout:Fetched 26.5 kB in 1s (27.3 kB/s) 2026-03-31T11:31:16.158 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:31:16.172 DEBUG:teuthology.orchestra.run.vm09:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-721-g5bb32787-1jammy cephadm=20.2.0-721-g5bb32787-1jammy ceph-mds=20.2.0-721-g5bb32787-1jammy ceph-mgr=20.2.0-721-g5bb32787-1jammy ceph-common=20.2.0-721-g5bb32787-1jammy ceph-fuse=20.2.0-721-g5bb32787-1jammy ceph-test=20.2.0-721-g5bb32787-1jammy ceph-volume=20.2.0-721-g5bb32787-1jammy radosgw=20.2.0-721-g5bb32787-1jammy python3-rados=20.2.0-721-g5bb32787-1jammy python3-rgw=20.2.0-721-g5bb32787-1jammy python3-cephfs=20.2.0-721-g5bb32787-1jammy python3-rbd=20.2.0-721-g5bb32787-1jammy libcephfs2=20.2.0-721-g5bb32787-1jammy libcephfs-dev=20.2.0-721-g5bb32787-1jammy librados2=20.2.0-721-g5bb32787-1jammy librbd1=20.2.0-721-g5bb32787-1jammy rbd-fuse=20.2.0-721-g5bb32787-1jammy 2026-03-31T11:31:16.173 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:31:16.186 DEBUG:teuthology.orchestra.run.vm06:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-721-g5bb32787-1jammy cephadm=20.2.0-721-g5bb32787-1jammy ceph-mds=20.2.0-721-g5bb32787-1jammy ceph-mgr=20.2.0-721-g5bb32787-1jammy ceph-common=20.2.0-721-g5bb32787-1jammy ceph-fuse=20.2.0-721-g5bb32787-1jammy ceph-test=20.2.0-721-g5bb32787-1jammy ceph-volume=20.2.0-721-g5bb32787-1jammy radosgw=20.2.0-721-g5bb32787-1jammy python3-rados=20.2.0-721-g5bb32787-1jammy python3-rgw=20.2.0-721-g5bb32787-1jammy python3-cephfs=20.2.0-721-g5bb32787-1jammy python3-rbd=20.2.0-721-g5bb32787-1jammy libcephfs2=20.2.0-721-g5bb32787-1jammy libcephfs-dev=20.2.0-721-g5bb32787-1jammy librados2=20.2.0-721-g5bb32787-1jammy librbd1=20.2.0-721-g5bb32787-1jammy rbd-fuse=20.2.0-721-g5bb32787-1jammy 2026-03-31T11:31:16.206 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:31:16.219 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:31:16.393 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:31:16.394 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:31:16.407 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:31:16.407 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:31:16.555 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:31:16.555 INFO:teuthology.orchestra.run.vm09.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:31:16.555 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-31T11:31:16.555 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout:The following additional packages will be installed: 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:31:16.556 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools socat xmlstarlet 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout:Suggested packages: 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout: mailx | mailutils 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout:Recommended packages: 2026-03-31T11:31:16.557 INFO:teuthology.orchestra.run.vm09.stdout: btrfs-tools 2026-03-31T11:31:16.600 INFO:teuthology.orchestra.run.vm09.stdout:The following NEW packages will be installed: 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm06.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-31T11:31:16.601 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm09.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm09.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm09.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools socat xmlstarlet 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout:The following additional packages will be installed: 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:31:16.602 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools socat xmlstarlet 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be upgraded: 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout:Suggested packages: 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout: mailx | mailutils 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout:Recommended packages: 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm06.stdout: btrfs-tools 2026-03-31T11:31:16.603 INFO:teuthology.orchestra.run.vm09.stdout: librados2 librbd1 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout:The following NEW packages will be installed: 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools socat xmlstarlet 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be upgraded: 2026-03-31T11:31:16.647 INFO:teuthology.orchestra.run.vm06.stdout: librados2 librbd1 2026-03-31T11:31:16.688 INFO:teuthology.orchestra.run.vm09.stdout:2 upgraded, 85 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:31:16.688 INFO:teuthology.orchestra.run.vm09.stdout:Need to get 281 MB of archives. 2026-03-31T11:31:16.688 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-31T11:31:16.688 INFO:teuthology.orchestra.run.vm09.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-31T11:31:16.757 INFO:teuthology.orchestra.run.vm06.stdout:2 upgraded, 85 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:31:16.757 INFO:teuthology.orchestra.run.vm06.stdout:Need to get 281 MB of archives. 2026-03-31T11:31:16.757 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-31T11:31:16.757 INFO:teuthology.orchestra.run.vm06.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-31T11:31:16.863 INFO:teuthology.orchestra.run.vm09.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-31T11:31:16.868 INFO:teuthology.orchestra.run.vm09.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-31T11:31:16.902 INFO:teuthology.orchestra.run.vm09.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-31T11:31:17.002 INFO:teuthology.orchestra.run.vm09.stdout:Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-31T11:31:17.003 INFO:teuthology.orchestra.run.vm06.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-31T11:31:17.006 INFO:teuthology.orchestra.run.vm09.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-31T11:31:17.019 INFO:teuthology.orchestra.run.vm09.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-31T11:31:17.023 INFO:teuthology.orchestra.run.vm09.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-31T11:31:17.024 INFO:teuthology.orchestra.run.vm09.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-31T11:31:17.024 INFO:teuthology.orchestra.run.vm09.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-31T11:31:17.025 INFO:teuthology.orchestra.run.vm09.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-31T11:31:17.033 INFO:teuthology.orchestra.run.vm09.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-31T11:31:17.033 INFO:teuthology.orchestra.run.vm09.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-31T11:31:17.036 INFO:teuthology.orchestra.run.vm09.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.2 [72.1 kB] 2026-03-31T11:31:17.061 INFO:teuthology.orchestra.run.vm06.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-31T11:31:17.062 INFO:teuthology.orchestra.run.vm06.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-31T11:31:17.068 INFO:teuthology.orchestra.run.vm09.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-31T11:31:17.069 INFO:teuthology.orchestra.run.vm09.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-31T11:31:17.069 INFO:teuthology.orchestra.run.vm09.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-31T11:31:17.069 INFO:teuthology.orchestra.run.vm09.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-31T11:31:17.069 INFO:teuthology.orchestra.run.vm09.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-31T11:31:17.070 INFO:teuthology.orchestra.run.vm09.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-31T11:31:17.070 INFO:teuthology.orchestra.run.vm09.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-31T11:31:17.071 INFO:teuthology.orchestra.run.vm09.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-31T11:31:17.072 INFO:teuthology.orchestra.run.vm09.stdout:Get:23 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-31T11:31:17.107 INFO:teuthology.orchestra.run.vm09.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-31T11:31:17.108 INFO:teuthology.orchestra.run.vm09.stdout:Get:25 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-31T11:31:17.108 INFO:teuthology.orchestra.run.vm09.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-31T11:31:17.109 INFO:teuthology.orchestra.run.vm09.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-31T11:31:17.161 INFO:teuthology.orchestra.run.vm09.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-31T11:31:17.162 INFO:teuthology.orchestra.run.vm09.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-31T11:31:17.162 INFO:teuthology.orchestra.run.vm09.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-31T11:31:17.181 INFO:teuthology.orchestra.run.vm09.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-31T11:31:17.181 INFO:teuthology.orchestra.run.vm09.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-31T11:31:17.181 INFO:teuthology.orchestra.run.vm09.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-31T11:31:17.181 INFO:teuthology.orchestra.run.vm09.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-31T11:31:17.181 INFO:teuthology.orchestra.run.vm09.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-31T11:31:17.182 INFO:teuthology.orchestra.run.vm09.stdout:Get:36 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-31T11:31:17.211 INFO:teuthology.orchestra.run.vm09.stdout:Get:37 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-31T11:31:17.212 INFO:teuthology.orchestra.run.vm09.stdout:Get:38 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-31T11:31:17.213 INFO:teuthology.orchestra.run.vm09.stdout:Get:39 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-31T11:31:17.238 INFO:teuthology.orchestra.run.vm06.stdout:Get:5 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-721-g5bb32787-1jammy [2867 kB] 2026-03-31T11:31:17.238 INFO:teuthology.orchestra.run.vm06.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-31T11:31:17.238 INFO:teuthology.orchestra.run.vm06.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-31T11:31:17.245 INFO:teuthology.orchestra.run.vm09.stdout:Get:40 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-31T11:31:17.248 INFO:teuthology.orchestra.run.vm09.stdout:Get:41 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-721-g5bb32787-1jammy [2867 kB] 2026-03-31T11:31:17.249 INFO:teuthology.orchestra.run.vm09.stdout:Get:42 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-31T11:31:17.251 INFO:teuthology.orchestra.run.vm09.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-31T11:31:17.256 INFO:teuthology.orchestra.run.vm09.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-31T11:31:17.293 INFO:teuthology.orchestra.run.vm09.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-31T11:31:17.293 INFO:teuthology.orchestra.run.vm09.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-31T11:31:17.294 INFO:teuthology.orchestra.run.vm09.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-31T11:31:17.310 INFO:teuthology.orchestra.run.vm06.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-31T11:31:17.315 INFO:teuthology.orchestra.run.vm09.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-31T11:31:17.315 INFO:teuthology.orchestra.run.vm09.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-31T11:31:17.318 INFO:teuthology.orchestra.run.vm06.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-31T11:31:17.318 INFO:teuthology.orchestra.run.vm09.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-31T11:31:17.318 INFO:teuthology.orchestra.run.vm09.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-31T11:31:17.319 INFO:teuthology.orchestra.run.vm06.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-31T11:31:17.319 INFO:teuthology.orchestra.run.vm06.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-31T11:31:17.319 INFO:teuthology.orchestra.run.vm06.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-31T11:31:17.319 INFO:teuthology.orchestra.run.vm09.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-31T11:31:17.320 INFO:teuthology.orchestra.run.vm09.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-31T11:31:17.322 INFO:teuthology.orchestra.run.vm06.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-31T11:31:17.322 INFO:teuthology.orchestra.run.vm06.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-31T11:31:17.322 INFO:teuthology.orchestra.run.vm06.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.2 [72.1 kB] 2026-03-31T11:31:17.339 INFO:teuthology.orchestra.run.vm06.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-31T11:31:17.350 INFO:teuthology.orchestra.run.vm09.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-31T11:31:17.372 INFO:teuthology.orchestra.run.vm06.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-31T11:31:17.372 INFO:teuthology.orchestra.run.vm06.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-31T11:31:17.372 INFO:teuthology.orchestra.run.vm06.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-31T11:31:17.372 INFO:teuthology.orchestra.run.vm06.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-31T11:31:17.372 INFO:teuthology.orchestra.run.vm06.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-31T11:31:17.372 INFO:teuthology.orchestra.run.vm06.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-31T11:31:17.375 INFO:teuthology.orchestra.run.vm06.stdout:Get:23 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-31T11:31:17.375 INFO:teuthology.orchestra.run.vm06.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-31T11:31:17.387 INFO:teuthology.orchestra.run.vm06.stdout:Get:25 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-31T11:31:17.400 INFO:teuthology.orchestra.run.vm09.stdout:Get:55 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-31T11:31:17.436 INFO:teuthology.orchestra.run.vm06.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-31T11:31:17.436 INFO:teuthology.orchestra.run.vm06.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-31T11:31:17.436 INFO:teuthology.orchestra.run.vm06.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-31T11:31:17.509 INFO:teuthology.orchestra.run.vm06.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-31T11:31:17.509 INFO:teuthology.orchestra.run.vm06.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-31T11:31:17.509 INFO:teuthology.orchestra.run.vm06.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-31T11:31:17.618 INFO:teuthology.orchestra.run.vm06.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-31T11:31:17.618 INFO:teuthology.orchestra.run.vm06.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-31T11:31:17.618 INFO:teuthology.orchestra.run.vm06.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-31T11:31:17.618 INFO:teuthology.orchestra.run.vm06.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-31T11:31:17.618 INFO:teuthology.orchestra.run.vm06.stdout:Get:36 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-31T11:31:17.619 INFO:teuthology.orchestra.run.vm06.stdout:Get:37 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-31T11:31:17.620 INFO:teuthology.orchestra.run.vm06.stdout:Get:38 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-31T11:31:17.621 INFO:teuthology.orchestra.run.vm06.stdout:Get:39 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-31T11:31:17.621 INFO:teuthology.orchestra.run.vm06.stdout:Get:40 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-31T11:31:17.667 INFO:teuthology.orchestra.run.vm06.stdout:Get:41 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-31T11:31:17.670 INFO:teuthology.orchestra.run.vm06.stdout:Get:42 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-31T11:31:17.715 INFO:teuthology.orchestra.run.vm06.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-31T11:31:17.719 INFO:teuthology.orchestra.run.vm06.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-31T11:31:17.800 INFO:teuthology.orchestra.run.vm06.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-31T11:31:17.800 INFO:teuthology.orchestra.run.vm06.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-31T11:31:17.800 INFO:teuthology.orchestra.run.vm06.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-31T11:31:17.800 INFO:teuthology.orchestra.run.vm06.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-31T11:31:17.801 INFO:teuthology.orchestra.run.vm06.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-31T11:31:17.816 INFO:teuthology.orchestra.run.vm06.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-31T11:31:17.816 INFO:teuthology.orchestra.run.vm06.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-31T11:31:17.817 INFO:teuthology.orchestra.run.vm06.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-31T11:31:17.817 INFO:teuthology.orchestra.run.vm06.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-31T11:31:17.887 INFO:teuthology.orchestra.run.vm06.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-31T11:31:17.948 INFO:teuthology.orchestra.run.vm06.stdout:Get:55 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-31T11:31:18.170 INFO:teuthology.orchestra.run.vm06.stdout:Get:56 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-721-g5bb32787-1jammy [3571 kB] 2026-03-31T11:31:18.327 INFO:teuthology.orchestra.run.vm09.stdout:Get:56 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-721-g5bb32787-1jammy [3571 kB] 2026-03-31T11:31:18.732 INFO:teuthology.orchestra.run.vm06.stdout:Get:57 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-721-g5bb32787-1jammy [831 kB] 2026-03-31T11:31:18.931 INFO:teuthology.orchestra.run.vm09.stdout:Get:57 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-721-g5bb32787-1jammy [831 kB] 2026-03-31T11:31:18.949 INFO:teuthology.orchestra.run.vm06.stdout:Get:58 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-721-g5bb32787-1jammy [364 kB] 2026-03-31T11:31:18.953 INFO:teuthology.orchestra.run.vm06.stdout:Get:59 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-721-g5bb32787-1jammy [32.9 kB] 2026-03-31T11:31:18.953 INFO:teuthology.orchestra.run.vm06.stdout:Get:60 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-721-g5bb32787-1jammy [184 kB] 2026-03-31T11:31:18.966 INFO:teuthology.orchestra.run.vm06.stdout:Get:61 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-721-g5bb32787-1jammy [83.9 kB] 2026-03-31T11:31:19.061 INFO:teuthology.orchestra.run.vm06.stdout:Get:62 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-721-g5bb32787-1jammy [341 kB] 2026-03-31T11:31:19.064 INFO:teuthology.orchestra.run.vm06.stdout:Get:63 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-721-g5bb32787-1jammy [8696 kB] 2026-03-31T11:31:19.158 INFO:teuthology.orchestra.run.vm09.stdout:Get:58 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-721-g5bb32787-1jammy [364 kB] 2026-03-31T11:31:19.168 INFO:teuthology.orchestra.run.vm09.stdout:Get:59 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-721-g5bb32787-1jammy [32.9 kB] 2026-03-31T11:31:19.169 INFO:teuthology.orchestra.run.vm09.stdout:Get:60 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-721-g5bb32787-1jammy [184 kB] 2026-03-31T11:31:19.170 INFO:teuthology.orchestra.run.vm09.stdout:Get:61 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-721-g5bb32787-1jammy [83.9 kB] 2026-03-31T11:31:19.276 INFO:teuthology.orchestra.run.vm09.stdout:Get:62 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-721-g5bb32787-1jammy [341 kB] 2026-03-31T11:31:19.287 INFO:teuthology.orchestra.run.vm09.stdout:Get:63 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-721-g5bb32787-1jammy [8696 kB] 2026-03-31T11:31:20.623 INFO:teuthology.orchestra.run.vm06.stdout:Get:64 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-721-g5bb32787-1jammy [112 kB] 2026-03-31T11:31:20.623 INFO:teuthology.orchestra.run.vm06.stdout:Get:65 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-721-g5bb32787-1jammy [261 kB] 2026-03-31T11:31:20.633 INFO:teuthology.orchestra.run.vm06.stdout:Get:66 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-721-g5bb32787-1jammy [29.3 MB] 2026-03-31T11:31:20.712 INFO:teuthology.orchestra.run.vm09.stdout:Get:64 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-721-g5bb32787-1jammy [112 kB] 2026-03-31T11:31:20.712 INFO:teuthology.orchestra.run.vm09.stdout:Get:65 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-721-g5bb32787-1jammy [261 kB] 2026-03-31T11:31:20.819 INFO:teuthology.orchestra.run.vm09.stdout:Get:66 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-721-g5bb32787-1jammy [29.3 MB] 2026-03-31T11:31:25.113 INFO:teuthology.orchestra.run.vm09.stdout:Get:67 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-721-g5bb32787-1jammy [5416 kB] 2026-03-31T11:31:25.533 INFO:teuthology.orchestra.run.vm06.stdout:Get:67 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-721-g5bb32787-1jammy [5416 kB] 2026-03-31T11:31:25.923 INFO:teuthology.orchestra.run.vm09.stdout:Get:68 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-721-g5bb32787-1jammy [246 kB] 2026-03-31T11:31:25.930 INFO:teuthology.orchestra.run.vm09.stdout:Get:69 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-721-g5bb32787-1jammy [124 kB] 2026-03-31T11:31:25.931 INFO:teuthology.orchestra.run.vm09.stdout:Get:70 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-721-g5bb32787-1jammy [907 kB] 2026-03-31T11:31:26.064 INFO:teuthology.orchestra.run.vm09.stdout:Get:71 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-721-g5bb32787-1jammy [6393 kB] 2026-03-31T11:31:26.531 INFO:teuthology.orchestra.run.vm06.stdout:Get:68 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-721-g5bb32787-1jammy [246 kB] 2026-03-31T11:31:26.637 INFO:teuthology.orchestra.run.vm06.stdout:Get:69 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-721-g5bb32787-1jammy [124 kB] 2026-03-31T11:31:26.638 INFO:teuthology.orchestra.run.vm06.stdout:Get:70 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-721-g5bb32787-1jammy [907 kB] 2026-03-31T11:31:26.751 INFO:teuthology.orchestra.run.vm06.stdout:Get:71 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-721-g5bb32787-1jammy [6393 kB] 2026-03-31T11:31:26.998 INFO:teuthology.orchestra.run.vm09.stdout:Get:72 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-721-g5bb32787-1jammy [21.7 MB] 2026-03-31T11:31:27.642 INFO:teuthology.orchestra.run.vm06.stdout:Get:72 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-721-g5bb32787-1jammy [21.7 MB] 2026-03-31T11:31:29.840 INFO:teuthology.orchestra.run.vm09.stdout:Get:73 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-721-g5bb32787-1jammy [14.1 kB] 2026-03-31T11:31:29.840 INFO:teuthology.orchestra.run.vm09.stdout:Get:74 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-721-g5bb32787-1jammy [955 kB] 2026-03-31T11:31:29.960 INFO:teuthology.orchestra.run.vm09.stdout:Get:75 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-721-g5bb32787-1jammy [2341 kB] 2026-03-31T11:31:30.216 INFO:teuthology.orchestra.run.vm09.stdout:Get:76 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-721-g5bb32787-1jammy [1049 kB] 2026-03-31T11:31:30.222 INFO:teuthology.orchestra.run.vm06.stdout:Get:73 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-721-g5bb32787-1jammy [14.1 kB] 2026-03-31T11:31:30.223 INFO:teuthology.orchestra.run.vm06.stdout:Get:74 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-721-g5bb32787-1jammy [955 kB] 2026-03-31T11:31:30.312 INFO:teuthology.orchestra.run.vm06.stdout:Get:75 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-721-g5bb32787-1jammy [2341 kB] 2026-03-31T11:31:30.336 INFO:teuthology.orchestra.run.vm09.stdout:Get:77 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-721-g5bb32787-1jammy [179 kB] 2026-03-31T11:31:30.353 INFO:teuthology.orchestra.run.vm09.stdout:Get:78 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-721-g5bb32787-1jammy [45.5 MB] 2026-03-31T11:31:30.536 INFO:teuthology.orchestra.run.vm06.stdout:Get:76 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-721-g5bb32787-1jammy [1049 kB] 2026-03-31T11:31:30.645 INFO:teuthology.orchestra.run.vm06.stdout:Get:77 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-721-g5bb32787-1jammy [179 kB] 2026-03-31T11:31:30.647 INFO:teuthology.orchestra.run.vm06.stdout:Get:78 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-721-g5bb32787-1jammy [45.5 MB] 2026-03-31T11:31:35.249 INFO:teuthology.orchestra.run.vm06.stdout:Get:79 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-721-g5bb32787-1jammy [8625 kB] 2026-03-31T11:31:36.318 INFO:teuthology.orchestra.run.vm06.stdout:Get:80 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-721-g5bb32787-1jammy [14.2 kB] 2026-03-31T11:31:36.318 INFO:teuthology.orchestra.run.vm06.stdout:Get:81 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-721-g5bb32787-1jammy [99.5 MB] 2026-03-31T11:31:36.444 INFO:teuthology.orchestra.run.vm09.stdout:Get:79 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-721-g5bb32787-1jammy [8625 kB] 2026-03-31T11:31:37.619 INFO:teuthology.orchestra.run.vm09.stdout:Get:80 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-721-g5bb32787-1jammy [14.2 kB] 2026-03-31T11:31:37.619 INFO:teuthology.orchestra.run.vm09.stdout:Get:81 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-721-g5bb32787-1jammy [99.5 MB] 2026-03-31T11:31:50.899 INFO:teuthology.orchestra.run.vm06.stdout:Get:82 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-721-g5bb32787-1jammy [135 kB] 2026-03-31T11:31:50.899 INFO:teuthology.orchestra.run.vm06.stdout:Get:83 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-721-g5bb32787-1jammy [43.2 kB] 2026-03-31T11:31:50.900 INFO:teuthology.orchestra.run.vm06.stdout:Get:84 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-721-g5bb32787-1jammy [30.7 kB] 2026-03-31T11:31:50.900 INFO:teuthology.orchestra.run.vm06.stdout:Get:85 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-721-g5bb32787-1jammy [41.4 kB] 2026-03-31T11:31:50.900 INFO:teuthology.orchestra.run.vm06.stdout:Get:86 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-721-g5bb32787-1jammy [25.1 MB] 2026-03-31T11:31:54.292 INFO:teuthology.orchestra.run.vm06.stdout:Get:87 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-721-g5bb32787-1jammy [97.5 kB] 2026-03-31T11:31:54.548 INFO:teuthology.orchestra.run.vm06.stdout:Fetched 281 MB in 38s (7454 kB/s) 2026-03-31T11:31:54.613 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-31T11:31:54.642 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-31T11:31:54.643 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-31T11:31:54.645 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-31T11:31:54.665 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-31T11:31:54.672 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-31T11:31:54.673 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-31T11:31:54.688 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-31T11:31:54.694 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-31T11:31:54.695 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-31T11:31:54.714 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-31T11:31:54.720 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-31T11:31:54.724 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:31:54.765 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-31T11:31:54.771 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-31T11:31:54.772 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:31:54.790 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-31T11:31:54.795 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-31T11:31:54.796 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:31:54.818 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-31T11:31:54.823 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-31T11:31:54.824 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-31T11:31:54.847 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../07-librbd1_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:54.849 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking librbd1 (20.2.0-721-g5bb32787-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-31T11:31:54.904 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../08-librados2_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:54.906 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking librados2 (20.2.0-721-g5bb32787-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-31T11:31:54.956 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libnbd0. 2026-03-31T11:31:54.962 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-31T11:31:54.963 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-31T11:31:54.978 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libcephfs2. 2026-03-31T11:31:54.983 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:54.984 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libcephfs2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.005 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-rados. 2026-03-31T11:31:55.010 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../11-python3-rados_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:55.011 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-rados (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.029 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-31T11:31:55.034 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:31:55.035 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-ceph-argparse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.047 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-cephfs. 2026-03-31T11:31:55.053 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:55.054 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-cephfs (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.069 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-31T11:31:55.075 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:31:55.076 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.096 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-31T11:31:55.102 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-31T11:31:55.103 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-31T11:31:55.120 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-prettytable. 2026-03-31T11:31:55.126 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-31T11:31:55.127 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-31T11:31:55.142 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-rbd. 2026-03-31T11:31:55.149 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:55.150 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-rbd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.169 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-31T11:31:55.174 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-31T11:31:55.175 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-31T11:31:55.196 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package librgw2. 2026-03-31T11:31:55.202 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../19-librgw2_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:55.203 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking librgw2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.321 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-rgw. 2026-03-31T11:31:55.326 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:55.327 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-rgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.344 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-31T11:31:55.349 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-31T11:31:55.350 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-31T11:31:55.365 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libradosstriper1. 2026-03-31T11:31:55.370 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:55.371 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libradosstriper1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.389 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-common. 2026-03-31T11:31:55.395 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../23-ceph-common_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:55.395 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.733 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-base. 2026-03-31T11:31:55.738 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../24-ceph-base_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:55.743 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-base (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:55.824 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-31T11:31:55.830 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-31T11:31:55.830 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-31T11:31:55.845 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-cheroot. 2026-03-31T11:31:55.851 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.2_all.deb ... 2026-03-31T11:31:55.851 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.2) ... 2026-03-31T11:31:55.869 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-31T11:31:55.876 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-31T11:31:55.877 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-31T11:31:55.893 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-31T11:31:55.898 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-31T11:31:55.899 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-31T11:31:55.914 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-31T11:31:55.919 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-31T11:31:55.920 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-31T11:31:55.934 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-tempora. 2026-03-31T11:31:55.940 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-31T11:31:55.940 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-31T11:31:55.955 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-portend. 2026-03-31T11:31:55.961 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-31T11:31:55.961 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-31T11:31:55.975 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-31T11:31:55.981 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-31T11:31:55.981 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-31T11:31:55.995 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-31T11:31:56.001 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-31T11:31:56.002 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-31T11:31:56.030 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-natsort. 2026-03-31T11:31:56.037 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-31T11:31:56.037 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-31T11:31:56.053 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-31T11:31:56.059 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:31:56.060 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-modules-core (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.091 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-31T11:31:56.097 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:56.098 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.113 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr. 2026-03-31T11:31:56.119 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:56.120 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.143 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mon. 2026-03-31T11:31:56.149 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:56.150 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.209 INFO:teuthology.orchestra.run.vm09.stdout:Get:82 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-721-g5bb32787-1jammy [135 kB] 2026-03-31T11:31:56.230 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-31T11:31:56.236 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-31T11:31:56.237 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-31T11:31:56.257 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-osd. 2026-03-31T11:31:56.261 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:56.262 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-osd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.267 INFO:teuthology.orchestra.run.vm09.stdout:Get:83 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-721-g5bb32787-1jammy [43.2 kB] 2026-03-31T11:31:56.267 INFO:teuthology.orchestra.run.vm09.stdout:Get:84 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-721-g5bb32787-1jammy [30.7 kB] 2026-03-31T11:31:56.267 INFO:teuthology.orchestra.run.vm09.stdout:Get:85 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-721-g5bb32787-1jammy [41.4 kB] 2026-03-31T11:31:56.270 INFO:teuthology.orchestra.run.vm09.stdout:Get:86 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-721-g5bb32787-1jammy [25.1 MB] 2026-03-31T11:31:56.466 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph. 2026-03-31T11:31:56.471 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../41-ceph_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:56.472 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.486 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-fuse. 2026-03-31T11:31:56.492 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:56.493 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.519 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mds. 2026-03-31T11:31:56.526 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:56.527 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mds (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.564 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package cephadm. 2026-03-31T11:31:56.570 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../44-cephadm_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:56.570 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.588 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-31T11:31:56.593 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-31T11:31:56.594 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-31T11:31:56.618 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-31T11:31:56.623 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:31:56.624 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:56.647 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-31T11:31:56.652 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-31T11:31:56.653 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-31T11:31:56.667 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-routes. 2026-03-31T11:31:56.673 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-31T11:31:56.673 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-31T11:31:56.695 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-31T11:31:56.700 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:31:56.701 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-dashboard (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:57.215 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-31T11:31:57.222 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-31T11:31:57.223 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-31T11:31:57.272 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-joblib. 2026-03-31T11:31:57.275 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-31T11:31:57.276 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-31T11:31:57.307 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-31T11:31:57.312 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-31T11:31:57.313 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-31T11:31:57.328 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-sklearn. 2026-03-31T11:31:57.333 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-31T11:31:57.334 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-31T11:31:57.439 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-31T11:31:57.445 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:31:57.446 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:57.638 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-cachetools. 2026-03-31T11:31:57.644 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-31T11:31:57.645 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-31T11:31:57.659 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-rsa. 2026-03-31T11:31:57.665 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-31T11:31:57.666 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-31T11:31:57.683 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-google-auth. 2026-03-31T11:31:57.689 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-31T11:31:57.690 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-31T11:31:57.708 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-31T11:31:57.714 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-31T11:31:57.715 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-31T11:31:57.730 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-websocket. 2026-03-31T11:31:57.736 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-31T11:31:57.737 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-31T11:31:57.755 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-31T11:31:57.761 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-31T11:31:57.762 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-31T11:31:57.886 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-31T11:31:57.892 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:31:57.893 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:57.907 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-31T11:31:57.913 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-31T11:31:57.914 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-31T11:31:57.930 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-31T11:31:57.936 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-31T11:31:57.937 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-31T11:31:57.950 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package jq. 2026-03-31T11:31:57.956 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-31T11:31:57.957 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-31T11:31:57.969 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package socat. 2026-03-31T11:31:57.975 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-31T11:31:57.976 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-31T11:31:57.996 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package xmlstarlet. 2026-03-31T11:31:58.002 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-31T11:31:58.003 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-31T11:31:58.045 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-test. 2026-03-31T11:31:58.051 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../67-ceph-test_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:58.051 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-test (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:58.965 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-volume. 2026-03-31T11:31:58.972 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:31:58.972 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-volume (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:59.000 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-31T11:31:59.004 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:59.004 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libcephfs-daemon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:59.018 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-31T11:31:59.024 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:59.025 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libcephfs-proxy2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:59.038 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-31T11:31:59.044 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:59.045 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libcephfs-dev (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:59.099 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package nvme-cli. 2026-03-31T11:31:59.105 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-31T11:31:59.106 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-31T11:31:59.144 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-31T11:31:59.150 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-31T11:31:59.151 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-31T11:31:59.186 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-31T11:31:59.192 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-31T11:31:59.193 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-31T11:31:59.206 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-pluggy. 2026-03-31T11:31:59.212 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-31T11:31:59.213 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-31T11:31:59.227 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-psutil. 2026-03-31T11:31:59.233 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-31T11:31:59.234 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-31T11:31:59.253 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-py. 2026-03-31T11:31:59.258 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-31T11:31:59.259 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-31T11:31:59.279 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-pygments. 2026-03-31T11:31:59.285 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-31T11:31:59.286 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-31T11:31:59.339 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-toml. 2026-03-31T11:31:59.345 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-31T11:31:59.345 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-31T11:31:59.364 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-pytest. 2026-03-31T11:31:59.366 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-31T11:31:59.367 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-31T11:31:59.404 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-simplejson. 2026-03-31T11:31:59.408 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-31T11:31:59.408 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-31T11:31:59.427 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-webob. 2026-03-31T11:31:59.433 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-31T11:31:59.433 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-31T11:31:59.451 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-31T11:31:59.457 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-31T11:31:59.457 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-31T11:31:59.547 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package radosgw. 2026-03-31T11:31:59.553 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../84-radosgw_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:59.554 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking radosgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:59.831 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package rbd-fuse. 2026-03-31T11:31:59.837 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:31:59.838 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking rbd-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:31:59.854 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package smartmontools. 2026-03-31T11:31:59.860 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-31T11:31:59.868 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-31T11:31:59.904 INFO:teuthology.orchestra.run.vm06.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-31T11:32:00.138 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-31T11:32:00.138 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-31T11:32:00.488 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-31T11:32:00.550 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-31T11:32:00.552 INFO:teuthology.orchestra.run.vm06.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-31T11:32:00.613 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-31T11:32:00.830 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-31T11:32:00.906 INFO:teuthology.orchestra.run.vm09.stdout:Get:87 https://3.chacra.ceph.com/r/ceph/tentacle-release/5bb3278730741031382ca9c3dc9d221a942e06a2/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-721-g5bb32787-1jammy [97.5 kB] 2026-03-31T11:32:01.162 INFO:teuthology.orchestra.run.vm09.stdout:Fetched 281 MB in 44s (6334 kB/s) 2026-03-31T11:32:01.170 INFO:teuthology.orchestra.run.vm06.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-31T11:32:01.188 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-31T11:32:01.191 INFO:teuthology.orchestra.run.vm06.stdout:Setting up cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.219 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-31T11:32:01.220 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-31T11:32:01.222 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-31T11:32:01.233 INFO:teuthology.orchestra.run.vm06.stdout:Adding system user cephadm....done 2026-03-31T11:32:01.241 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-31T11:32:01.242 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-31T11:32:01.247 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-31T11:32:01.248 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-31T11:32:01.263 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-31T11:32:01.270 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-31T11:32:01.271 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-31T11:32:01.291 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-31T11:32:01.297 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-31T11:32:01.302 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:01.310 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-31T11:32:01.312 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-31T11:32:01.335 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-31T11:32:01.341 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-31T11:32:01.342 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:01.360 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-31T11:32:01.366 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-31T11:32:01.366 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:01.378 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-31T11:32:01.389 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-31T11:32:01.394 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-31T11:32:01.395 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-31T11:32:01.419 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../07-librbd1_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.421 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking librbd1 (20.2.0-721-g5bb32787-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-31T11:32:01.448 INFO:teuthology.orchestra.run.vm06.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-31T11:32:01.450 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-31T11:32:01.479 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../08-librados2_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.481 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking librados2 (20.2.0-721-g5bb32787-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-31T11:32:01.535 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libnbd0. 2026-03-31T11:32:01.541 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-31T11:32:01.542 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-31T11:32:01.542 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-31T11:32:01.557 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libcephfs2. 2026-03-31T11:32:01.563 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.564 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libcephfs2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.587 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-rados. 2026-03-31T11:32:01.592 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../11-python3-rados_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.593 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-rados (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.612 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-31T11:32:01.617 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:32:01.618 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-ceph-argparse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.632 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-cephfs. 2026-03-31T11:32:01.637 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.638 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-cephfs (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.655 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-31T11:32:01.660 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-31T11:32:01.661 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:32:01.662 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.683 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-31T11:32:01.689 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-31T11:32:01.690 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-31T11:32:01.707 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-prettytable. 2026-03-31T11:32:01.712 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-31T11:32:01.713 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-31T11:32:01.728 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-rbd. 2026-03-31T11:32:01.729 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-31T11:32:01.733 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.734 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-rbd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.754 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-31T11:32:01.761 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-31T11:32:01.762 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-31T11:32:01.782 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package librgw2. 2026-03-31T11:32:01.788 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../19-librgw2_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.789 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking librgw2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.797 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-ceph-argparse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.871 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-31T11:32:01.873 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-31T11:32:01.875 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-31T11:32:01.878 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-31T11:32:01.886 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-31T11:32:01.914 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-rgw. 2026-03-31T11:32:01.920 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.921 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-rgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.938 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-31T11:32:01.945 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-31T11:32:01.946 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-31T11:32:01.960 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libradosstriper1. 2026-03-31T11:32:01.966 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.967 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libradosstriper1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:01.985 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-common. 2026-03-31T11:32:01.990 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../23-ceph-common_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:01.991 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.005 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-31T11:32:02.075 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libcephfs-proxy2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.077 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-31T11:32:02.148 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-31T11:32:02.227 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-31T11:32:02.323 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-base. 2026-03-31T11:32:02.329 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../24-ceph-base_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:02.333 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-base (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.416 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-31T11:32:02.423 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-31T11:32:02.424 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-31T11:32:02.439 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-cheroot. 2026-03-31T11:32:02.445 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.2_all.deb ... 2026-03-31T11:32:02.446 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.2) ... 2026-03-31T11:32:02.463 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-31T11:32:02.469 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-31T11:32:02.470 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-31T11:32:02.483 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-31T11:32:02.488 INFO:teuthology.orchestra.run.vm06.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-31T11:32:02.489 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-31T11:32:02.490 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-31T11:32:02.490 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-31T11:32:02.504 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-31T11:32:02.510 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-31T11:32:02.511 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-31T11:32:02.524 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-tempora. 2026-03-31T11:32:02.530 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-31T11:32:02.531 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-31T11:32:02.546 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-portend. 2026-03-31T11:32:02.552 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-31T11:32:02.553 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-31T11:32:02.567 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-31T11:32:02.573 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-31T11:32:02.574 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-31T11:32:02.583 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-31T11:32:02.588 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-31T11:32:02.594 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-31T11:32:02.595 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-31T11:32:02.621 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-natsort. 2026-03-31T11:32:02.627 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-31T11:32:02.628 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-31T11:32:02.643 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-31T11:32:02.648 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:32:02.649 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-modules-core (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.680 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-31T11:32:02.686 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:02.687 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.705 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr. 2026-03-31T11:32:02.711 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:02.711 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.721 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.2) ... 2026-03-31T11:32:02.735 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mon. 2026-03-31T11:32:02.741 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:02.742 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.807 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-31T11:32:02.816 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-31T11:32:02.821 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-31T11:32:02.822 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-31T11:32:02.840 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-osd. 2026-03-31T11:32:02.846 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:02.847 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-osd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.872 INFO:teuthology.orchestra.run.vm06.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-31T11:32:02.874 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:02.967 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-31T11:32:03.051 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph. 2026-03-31T11:32:03.057 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../41-ceph_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:03.058 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:03.074 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-fuse. 2026-03-31T11:32:03.080 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:03.081 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:03.106 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mds. 2026-03-31T11:32:03.112 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:03.113 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mds (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:03.150 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package cephadm. 2026-03-31T11:32:03.155 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../44-cephadm_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:03.156 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:03.174 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-31T11:32:03.180 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-31T11:32:03.181 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-31T11:32:03.205 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-31T11:32:03.211 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:32:03.212 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:03.235 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-31T11:32:03.241 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-31T11:32:03.242 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-31T11:32:03.257 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-routes. 2026-03-31T11:32:03.262 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-31T11:32:03.263 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-31T11:32:03.285 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-31T11:32:03.290 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:32:03.291 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-dashboard (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:03.498 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:03.504 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-31T11:32:03.577 INFO:teuthology.orchestra.run.vm06.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-31T11:32:03.579 INFO:teuthology.orchestra.run.vm06.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-31T11:32:03.581 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-31T11:32:03.654 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-31T11:32:03.720 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:03.722 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-31T11:32:03.790 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-31T11:32:03.800 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-31T11:32:03.806 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-31T11:32:03.807 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-31T11:32:03.855 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-joblib. 2026-03-31T11:32:03.856 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-31T11:32:03.861 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-31T11:32:03.862 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-31T11:32:03.896 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-31T11:32:03.902 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-31T11:32:03.903 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-31T11:32:03.919 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-sklearn. 2026-03-31T11:32:03.925 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-31T11:32:03.926 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-31T11:32:03.928 INFO:teuthology.orchestra.run.vm06.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-31T11:32:03.930 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-31T11:32:04.008 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-31T11:32:04.009 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-31T11:32:04.040 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-31T11:32:04.047 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:32:04.048 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:04.080 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-31T11:32:04.169 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-31T11:32:04.232 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-31T11:32:04.234 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-31T11:32:04.246 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-cachetools. 2026-03-31T11:32:04.252 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-31T11:32:04.253 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-31T11:32:04.268 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-rsa. 2026-03-31T11:32:04.274 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-31T11:32:04.275 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-31T11:32:04.293 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-google-auth. 2026-03-31T11:32:04.299 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-31T11:32:04.300 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-31T11:32:04.318 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-31T11:32:04.324 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-31T11:32:04.325 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-31T11:32:04.341 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-websocket. 2026-03-31T11:32:04.347 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-31T11:32:04.351 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-31T11:32:04.364 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-31T11:32:04.369 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-31T11:32:04.375 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-31T11:32:04.376 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-31T11:32:04.432 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:04.434 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-31T11:32:04.498 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-31T11:32:04.504 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:32:04.505 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:04.510 INFO:teuthology.orchestra.run.vm06.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-31T11:32:04.512 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-31T11:32:04.520 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-31T11:32:04.526 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-31T11:32:04.527 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-31T11:32:04.544 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-31T11:32:04.550 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-31T11:32:04.551 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-31T11:32:04.565 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package jq. 2026-03-31T11:32:04.571 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-31T11:32:04.572 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-31T11:32:04.585 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package socat. 2026-03-31T11:32:04.590 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-31T11:32:04.591 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-31T11:32:04.612 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package xmlstarlet. 2026-03-31T11:32:04.618 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-31T11:32:04.618 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-31T11:32:04.644 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-31T11:32:04.646 INFO:teuthology.orchestra.run.vm06.stdout:Setting up librados2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:04.648 INFO:teuthology.orchestra.run.vm06.stdout:Setting up librgw2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:04.649 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libsqlite3-mod-ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:04.652 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-31T11:32:04.658 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-test. 2026-03-31T11:32:04.664 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../67-ceph-test_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:04.665 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-test (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.199 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libcephfs2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.201 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libradosstriper1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.203 INFO:teuthology.orchestra.run.vm06.stdout:Setting up librbd1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.205 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-modules-core (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.207 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.271 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-31T11:32:05.271 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-31T11:32:05.601 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-volume. 2026-03-31T11:32:05.607 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-721-g5bb32787-1jammy_all.deb ... 2026-03-31T11:32:05.608 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-volume (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.614 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libcephfs-dev (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.615 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-rados (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.617 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libcephfs-daemon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.620 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-rbd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.621 INFO:teuthology.orchestra.run.vm06.stdout:Setting up rbd-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.623 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-rgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.625 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-cephfs (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.627 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.633 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-31T11:32:05.638 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:05.639 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libcephfs-daemon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.654 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-31T11:32:05.660 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:05.661 INFO:teuthology.orchestra.run.vm06.stdout:Adding group ceph....done 2026-03-31T11:32:05.661 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libcephfs-proxy2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.677 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-31T11:32:05.680 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:05.681 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libcephfs-dev (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:05.696 INFO:teuthology.orchestra.run.vm06.stdout:Adding system user ceph....done 2026-03-31T11:32:05.698 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package nvme-cli. 2026-03-31T11:32:05.703 INFO:teuthology.orchestra.run.vm06.stdout:Setting system user ceph properties....done 2026-03-31T11:32:05.704 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-31T11:32:05.705 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-31T11:32:05.707 INFO:teuthology.orchestra.run.vm06.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-31T11:32:05.738 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-31T11:32:05.744 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-31T11:32:05.745 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-31T11:32:05.771 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-31T11:32:05.780 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-31T11:32:05.786 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-31T11:32:05.787 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-31T11:32:05.802 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pluggy. 2026-03-31T11:32:05.808 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-31T11:32:05.809 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-31T11:32:05.825 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-psutil. 2026-03-31T11:32:05.831 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-31T11:32:05.832 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-31T11:32:05.853 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-py. 2026-03-31T11:32:05.859 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-31T11:32:05.860 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-31T11:32:05.882 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pygments. 2026-03-31T11:32:05.887 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-31T11:32:05.888 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-31T11:32:05.939 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-toml. 2026-03-31T11:32:05.946 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-31T11:32:05.946 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-31T11:32:05.962 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pytest. 2026-03-31T11:32:05.968 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-31T11:32:05.969 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-31T11:32:05.987 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-31T11:32:06.004 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-simplejson. 2026-03-31T11:32:06.010 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-31T11:32:06.011 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-31T11:32:06.030 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-webob. 2026-03-31T11:32:06.036 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-31T11:32:06.037 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-31T11:32:06.054 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-31T11:32:06.060 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-31T11:32:06.060 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-31T11:32:06.149 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package radosgw. 2026-03-31T11:32:06.155 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../84-radosgw_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:06.156 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking radosgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:06.366 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-test (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:06.368 INFO:teuthology.orchestra.run.vm06.stdout:Setting up radosgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:06.435 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package rbd-fuse. 2026-03-31T11:32:06.441 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-721-g5bb32787-1jammy_amd64.deb ... 2026-03-31T11:32:06.442 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking rbd-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:06.459 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package smartmontools. 2026-03-31T11:32:06.465 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-31T11:32:06.473 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-31T11:32:06.509 INFO:teuthology.orchestra.run.vm09.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-31T11:32:06.595 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-31T11:32:06.595 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-31T11:32:06.752 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-31T11:32:06.752 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-31T11:32:06.946 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-base (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:07.035 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-31T11:32:07.089 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-31T11:32:07.152 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-31T11:32:07.155 INFO:teuthology.orchestra.run.vm09.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-31T11:32:07.221 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-31T11:32:07.414 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mds (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:07.449 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-31T11:32:07.475 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-31T11:32:07.475 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-31T11:32:07.821 INFO:teuthology.orchestra.run.vm09.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-31T11:32:07.837 INFO:teuthology.orchestra.run.vm09.stdout:Setting up cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:07.838 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:07.876 INFO:teuthology.orchestra.run.vm09.stdout:Adding system user cephadm....done 2026-03-31T11:32:07.885 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-31T11:32:07.909 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-31T11:32:07.909 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-31T11:32:07.946 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-31T11:32:07.948 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-31T11:32:08.015 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-31T11:32:08.080 INFO:teuthology.orchestra.run.vm09.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-31T11:32:08.082 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-31T11:32:08.168 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-31T11:32:08.241 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-osd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:08.284 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-31T11:32:08.316 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-31T11:32:08.316 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-31T11:32:08.349 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-31T11:32:08.416 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-ceph-argparse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:08.481 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-31T11:32:08.483 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-31T11:32:08.486 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-31T11:32:08.488 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-31T11:32:08.490 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-31T11:32:08.604 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-31T11:32:08.672 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libcephfs-proxy2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:08.674 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-31T11:32:08.685 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-k8sevents (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:08.687 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:08.699 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:08.744 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-31T11:32:08.758 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-31T11:32:08.758 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-31T11:32:08.827 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-31T11:32:09.084 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:09.085 INFO:teuthology.orchestra.run.vm09.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-31T11:32:09.087 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-31T11:32:09.097 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:09.098 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-dashboard (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:09.110 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-volume (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:09.175 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-31T11:32:09.227 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:32:09.299 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:32:09.304 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.2) ... 2026-03-31T11:32:09.388 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-31T11:32:09.453 INFO:teuthology.orchestra.run.vm09.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-31T11:32:09.455 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:09.548 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-31T11:32:09.585 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:09.585 INFO:teuthology.orchestra.run.vm06.stdout:Running kernel seems to be up-to-date. 2026-03-31T11:32:09.585 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:09.585 INFO:teuthology.orchestra.run.vm06.stdout:Services to be restarted: 2026-03-31T11:32:09.588 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart apache-htcacheclean.service 2026-03-31T11:32:09.593 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart rsyslog.service 2026-03-31T11:32:09.596 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout:Service restarts being deferred: 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart networkd-dispatcher.service 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart unattended-upgrades.service 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout:No containers need to be restarted. 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout:No user sessions are running outdated binaries. 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:09.597 INFO:teuthology.orchestra.run.vm06.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-31T11:32:10.096 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:10.101 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-31T11:32:10.170 INFO:teuthology.orchestra.run.vm09.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-31T11:32:10.172 INFO:teuthology.orchestra.run.vm09.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-31T11:32:10.175 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-31T11:32:10.241 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-31T11:32:10.304 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:10.306 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-31T11:32:10.376 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-31T11:32:10.433 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:32:10.436 DEBUG:teuthology.orchestra.run.vm06:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-jmespath python3-xmltodict s3cmd 2026-03-31T11:32:10.443 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-31T11:32:10.510 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:32:10.513 INFO:teuthology.orchestra.run.vm09.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-31T11:32:10.515 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-31T11:32:10.589 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-31T11:32:10.591 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-31T11:32:10.659 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-31T11:32:10.685 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:32:10.685 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:32:10.751 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-31T11:32:10.813 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:32:10.813 INFO:teuthology.orchestra.run.vm06.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:32:10.814 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-31T11:32:10.814 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:32:10.817 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-31T11:32:10.820 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-31T11:32:10.828 INFO:teuthology.orchestra.run.vm06.stdout:The following NEW packages will be installed: 2026-03-31T11:32:10.828 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath python3-xmltodict s3cmd 2026-03-31T11:32:10.950 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-31T11:32:11.011 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:32:11.014 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-31T11:32:11.050 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 3 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:32:11.050 INFO:teuthology.orchestra.run.vm06.stdout:Need to get 155 kB of archives. 2026-03-31T11:32:11.050 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 678 kB of additional disk space will be used. 2026-03-31T11:32:11.050 INFO:teuthology.orchestra.run.vm06.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-31T11:32:11.088 INFO:teuthology.orchestra.run.vm09.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-31T11:32:11.090 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-31T11:32:11.213 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-31T11:32:11.215 INFO:teuthology.orchestra.run.vm09.stdout:Setting up librados2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:11.218 INFO:teuthology.orchestra.run.vm09.stdout:Setting up librgw2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:11.220 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libsqlite3-mod-ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:11.222 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-31T11:32:11.268 INFO:teuthology.orchestra.run.vm06.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-31T11:32:11.293 INFO:teuthology.orchestra.run.vm06.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-31T11:32:11.698 INFO:teuthology.orchestra.run.vm06.stdout:Fetched 155 kB in 1s (226 kB/s) 2026-03-31T11:32:11.710 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jmespath. 2026-03-31T11:32:11.739 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-31T11:32:11.741 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-31T11:32:11.743 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-31T11:32:11.749 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libcephfs2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:11.752 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libradosstriper1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:11.754 INFO:teuthology.orchestra.run.vm09.stdout:Setting up librbd1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:11.757 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-modules-core (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:11.759 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-31T11:32:11.759 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:11.766 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-31T11:32:11.767 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-31T11:32:11.783 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package s3cmd. 2026-03-31T11:32:11.789 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-31T11:32:11.789 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-31T11:32:11.818 INFO:teuthology.orchestra.run.vm06.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-31T11:32:11.820 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-31T11:32:11.820 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-31T11:32:11.902 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-31T11:32:11.964 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-31T11:32:12.034 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:32:12.175 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libcephfs-dev (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.177 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-rados (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.179 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libcephfs-daemon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.182 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-rbd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.184 INFO:teuthology.orchestra.run.vm09.stdout:Setting up rbd-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.186 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-rgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.188 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-cephfs (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.190 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.223 INFO:teuthology.orchestra.run.vm09.stdout:Adding group ceph....done 2026-03-31T11:32:12.258 INFO:teuthology.orchestra.run.vm09.stdout:Adding system user ceph....done 2026-03-31T11:32:12.267 INFO:teuthology.orchestra.run.vm09.stdout:Setting system user ceph properties....done 2026-03-31T11:32:12.271 INFO:teuthology.orchestra.run.vm09.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-31T11:32:12.333 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-31T11:32:12.340 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:12.340 INFO:teuthology.orchestra.run.vm06.stdout:Running kernel seems to be up-to-date. 2026-03-31T11:32:12.340 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:12.340 INFO:teuthology.orchestra.run.vm06.stdout:Services to be restarted: 2026-03-31T11:32:12.342 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart apache-htcacheclean.service 2026-03-31T11:32:12.348 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart rsyslog.service 2026-03-31T11:32:12.351 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout:Service restarts being deferred: 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart networkd-dispatcher.service 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart unattended-upgrades.service 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout:No containers need to be restarted. 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout:No user sessions are running outdated binaries. 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:32:12.352 INFO:teuthology.orchestra.run.vm06.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-31T11:32:12.570 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-31T11:32:12.932 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-test (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:12.934 INFO:teuthology.orchestra.run.vm09.stdout:Setting up radosgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:13.159 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-31T11:32:13.159 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-31T11:32:13.216 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:32:13.219 DEBUG:teuthology.parallel:result is None 2026-03-31T11:32:13.522 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-base (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:13.607 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-31T11:32:13.965 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mds (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:14.025 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-31T11:32:14.025 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-31T11:32:14.401 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:14.471 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-31T11:32:14.471 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-31T11:32:14.837 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-osd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:14.909 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-31T11:32:14.909 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-31T11:32:15.322 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-k8sevents (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:15.324 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:15.336 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:15.397 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-31T11:32:15.397 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-31T11:32:15.762 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:15.775 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:15.777 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-dashboard (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:15.789 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-volume (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:32:15.900 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:32:15.972 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:32:16.253 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:16.253 INFO:teuthology.orchestra.run.vm09.stdout:Running kernel seems to be up-to-date. 2026-03-31T11:32:16.253 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:16.253 INFO:teuthology.orchestra.run.vm09.stdout:Services to be restarted: 2026-03-31T11:32:16.256 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart apache-htcacheclean.service 2026-03-31T11:32:16.261 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart rsyslog.service 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout:Service restarts being deferred: 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart networkd-dispatcher.service 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart unattended-upgrades.service 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout:No containers need to be restarted. 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout:No user sessions are running outdated binaries. 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:16.264 INFO:teuthology.orchestra.run.vm09.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-31T11:32:17.055 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:32:17.057 DEBUG:teuthology.orchestra.run.vm09:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-jmespath python3-xmltodict s3cmd 2026-03-31T11:32:17.134 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:32:17.302 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:32:17.302 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:32:17.418 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:32:17.418 INFO:teuthology.orchestra.run.vm09.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:32:17.418 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-31T11:32:17.418 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:32:17.430 INFO:teuthology.orchestra.run.vm09.stdout:The following NEW packages will be installed: 2026-03-31T11:32:17.430 INFO:teuthology.orchestra.run.vm09.stdout: python3-jmespath python3-xmltodict s3cmd 2026-03-31T11:32:17.651 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 3 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:32:17.651 INFO:teuthology.orchestra.run.vm09.stdout:Need to get 155 kB of archives. 2026-03-31T11:32:17.651 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 678 kB of additional disk space will be used. 2026-03-31T11:32:17.651 INFO:teuthology.orchestra.run.vm09.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-31T11:32:17.881 INFO:teuthology.orchestra.run.vm09.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-31T11:32:17.907 INFO:teuthology.orchestra.run.vm09.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-31T11:32:18.321 INFO:teuthology.orchestra.run.vm09.stdout:Fetched 155 kB in 1s (216 kB/s) 2026-03-31T11:32:18.334 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jmespath. 2026-03-31T11:32:18.361 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-31T11:32:18.363 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-31T11:32:18.364 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-31T11:32:18.380 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-31T11:32:18.385 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-31T11:32:18.386 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-31T11:32:18.400 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package s3cmd. 2026-03-31T11:32:18.405 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-31T11:32:18.406 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-31T11:32:18.434 INFO:teuthology.orchestra.run.vm09.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-31T11:32:18.515 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-31T11:32:18.575 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-31T11:32:18.640 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:32:18.940 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:18.940 INFO:teuthology.orchestra.run.vm09.stdout:Running kernel seems to be up-to-date. 2026-03-31T11:32:18.940 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:18.940 INFO:teuthology.orchestra.run.vm09.stdout:Services to be restarted: 2026-03-31T11:32:18.942 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart apache-htcacheclean.service 2026-03-31T11:32:18.947 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart rsyslog.service 2026-03-31T11:32:18.949 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout:Service restarts being deferred: 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart networkd-dispatcher.service 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart unattended-upgrades.service 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout:No containers need to be restarted. 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout:No user sessions are running outdated binaries. 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:32:18.950 INFO:teuthology.orchestra.run.vm09.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-31T11:32:19.691 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:32:19.694 DEBUG:teuthology.parallel:result is None 2026-03-31T11:32:19.694 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=5bb3278730741031382ca9c3dc9d221a942e06a2 2026-03-31T11:32:20.265 DEBUG:teuthology.orchestra.run.vm06:> dpkg-query -W -f '${Version}' ceph 2026-03-31T11:32:20.274 INFO:teuthology.orchestra.run.vm06.stdout:20.2.0-721-g5bb32787-1jammy 2026-03-31T11:32:20.274 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-721-g5bb32787-1jammy 2026-03-31T11:32:20.274 INFO:teuthology.task.install:The correct ceph version 20.2.0-721-g5bb32787-1jammy is installed. 2026-03-31T11:32:20.274 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=5bb3278730741031382ca9c3dc9d221a942e06a2 2026-03-31T11:32:20.852 DEBUG:teuthology.orchestra.run.vm09:> dpkg-query -W -f '${Version}' ceph 2026-03-31T11:32:20.861 INFO:teuthology.orchestra.run.vm09.stdout:20.2.0-721-g5bb32787-1jammy 2026-03-31T11:32:20.861 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-721-g5bb32787-1jammy 2026-03-31T11:32:20.861 INFO:teuthology.task.install:The correct ceph version 20.2.0-721-g5bb32787-1jammy is installed. 2026-03-31T11:32:20.861 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-31T11:32:20.861 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:32:20.861 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-31T11:32:20.869 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:32:20.869 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-31T11:32:20.912 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-31T11:32:20.912 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:32:20.912 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-03-31T11:32:20.920 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-31T11:32:20.971 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:32:20.971 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/daemon-helper 2026-03-31T11:32:20.978 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-31T11:32:21.028 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-31T11:32:21.028 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:32:21.028 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-31T11:32:21.036 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-31T11:32:21.087 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:32:21.087 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-31T11:32:21.094 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-31T11:32:21.144 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-31T11:32:21.144 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:32:21.144 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-03-31T11:32:21.153 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-31T11:32:21.203 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:32:21.203 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/stdin-killer 2026-03-31T11:32:21.210 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-31T11:32:21.260 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-31T11:32:21.302 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 3}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': True}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': '5bb3278730741031382ca9c3dc9d221a942e06a2', 'cephadm_binary_url': 'https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm'} 2026-03-31T11:32:21.302 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 2026-03-31T11:32:21.302 INFO:tasks.cephadm:Cluster fsid is 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:32:21.302 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-31T11:32:21.302 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.106', 'mon.b': '192.168.123.109'} 2026-03-31T11:32:21.302 INFO:tasks.cephadm:First mon is mon.a on vm06 2026-03-31T11:32:21.302 INFO:tasks.cephadm:First mgr is a 2026-03-31T11:32:21.302 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-31T11:32:21.302 DEBUG:teuthology.orchestra.run.vm06:> sudo hostname $(hostname -s) 2026-03-31T11:32:21.311 DEBUG:teuthology.orchestra.run.vm09:> sudo hostname $(hostname -s) 2026-03-31T11:32:21.319 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm 2026-03-31T11:32:21.319 DEBUG:teuthology.orchestra.run.vm06:> curl --silent -L https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-31T11:32:22.501 INFO:teuthology.orchestra.run.vm06.stdout:-rw-rw-r-- 1 ubuntu ubuntu 1036391 Mar 31 11:32 /home/ubuntu/cephtest/cephadm 2026-03-31T11:32:22.501 DEBUG:teuthology.orchestra.run.vm09:> curl --silent -L https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-31T11:32:23.655 INFO:teuthology.orchestra.run.vm09.stdout:-rw-rw-r-- 1 ubuntu ubuntu 1036391 Mar 31 11:32 /home/ubuntu/cephtest/cephadm 2026-03-31T11:32:23.656 DEBUG:teuthology.orchestra.run.vm06:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-31T11:32:23.660 DEBUG:teuthology.orchestra.run.vm09:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-31T11:32:23.667 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 on all hosts... 2026-03-31T11:32:23.667 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 pull 2026-03-31T11:32:23.705 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 pull 2026-03-31T11:32:23.943 INFO:teuthology.orchestra.run.vm06.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2... 2026-03-31T11:32:23.945 INFO:teuthology.orchestra.run.vm09.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2... 2026-03-31T11:33:04.588 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-03-31T11:33:04.588 INFO:teuthology.orchestra.run.vm06.stdout: "ceph_version": "ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable)", 2026-03-31T11:33:04.588 INFO:teuthology.orchestra.run.vm06.stdout: "image_id": "1e58a3cbf9abfa7cd4c97d6122dfc897574d910096f68804997a3e0f45bc44f0", 2026-03-31T11:33:04.588 INFO:teuthology.orchestra.run.vm06.stdout: "repo_digests": [ 2026-03-31T11:33:04.588 INFO:teuthology.orchestra.run.vm06.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:02c8d616f8a7af1a26efca44a51de7761356c5cb66c69c789e66f798c27c8072" 2026-03-31T11:33:04.588 INFO:teuthology.orchestra.run.vm06.stdout: ] 2026-03-31T11:33:04.588 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-03-31T11:33:07.985 INFO:teuthology.orchestra.run.vm09.stdout:{ 2026-03-31T11:33:07.986 INFO:teuthology.orchestra.run.vm09.stdout: "ceph_version": "ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable)", 2026-03-31T11:33:07.986 INFO:teuthology.orchestra.run.vm09.stdout: "image_id": "1e58a3cbf9abfa7cd4c97d6122dfc897574d910096f68804997a3e0f45bc44f0", 2026-03-31T11:33:07.986 INFO:teuthology.orchestra.run.vm09.stdout: "repo_digests": [ 2026-03-31T11:33:07.986 INFO:teuthology.orchestra.run.vm09.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:02c8d616f8a7af1a26efca44a51de7761356c5cb66c69c789e66f798c27c8072" 2026-03-31T11:33:07.986 INFO:teuthology.orchestra.run.vm09.stdout: ] 2026-03-31T11:33:07.986 INFO:teuthology.orchestra.run.vm09.stdout:} 2026-03-31T11:33:08.003 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph 2026-03-31T11:33:08.011 DEBUG:teuthology.orchestra.run.vm09:> sudo mkdir -p /etc/ceph 2026-03-31T11:33:08.019 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /etc/ceph 2026-03-31T11:33:08.059 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 777 /etc/ceph 2026-03-31T11:33:08.068 INFO:tasks.cephadm:Writing seed config... 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [global] mon election default strategy = 3 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = True 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-31T11:33:08.069 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-31T11:33:08.069 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:33:08.069 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-31T11:33:08.103 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 48134446-2cf5-11f1-830d-4121aa8d72d5 mon election default strategy = 3 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd mclock skip benchmark = true osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = True [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-31T11:33:08.103 DEBUG:teuthology.orchestra.run.vm06:mon.a> sudo journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a.service 2026-03-31T11:33:08.145 DEBUG:teuthology.orchestra.run.vm06:mgr.a> sudo journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.a.service 2026-03-31T11:33:08.188 INFO:tasks.cephadm:Bootstrapping... 2026-03-31T11:33:08.189 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 -v bootstrap --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.106 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:08.474 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-31T11:33:08.474 INFO:teuthology.orchestra.run.vm06.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2', '-v', 'bootstrap', '--fsid', '48134446-2cf5-11f1-830d-4121aa8d72d5', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.106', '--skip-admin-label'] 2026-03-31T11:33:08.474 INFO:teuthology.orchestra.run.vm06.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-31T11:33:08.474 INFO:teuthology.orchestra.run.vm06.stdout:Verifying podman|docker is present... 2026-03-31T11:33:08.474 INFO:teuthology.orchestra.run.vm06.stdout:Verifying lvm2 is present... 2026-03-31T11:33:08.474 INFO:teuthology.orchestra.run.vm06.stdout:Verifying time synchronization is in place... 2026-03-31T11:33:08.477 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-31T11:33:08.477 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-31T11:33:08.479 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-31T11:33:08.479 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:08.482 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-31T11:33:08.482 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-31T11:33:08.484 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-31T11:33:08.484 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:08.486 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-31T11:33:08.486 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout masked 2026-03-31T11:33:08.488 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-31T11:33:08.488 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:08.490 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-31T11:33:08.490 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-31T11:33:08.492 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-31T11:33:08.492 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:08.494 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout enabled 2026-03-31T11:33:08.496 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout active 2026-03-31T11:33:08.496 INFO:teuthology.orchestra.run.vm06.stdout:Unit ntp.service is enabled and running 2026-03-31T11:33:08.496 INFO:teuthology.orchestra.run.vm06.stdout:Repeating the final host check... 2026-03-31T11:33:08.496 INFO:teuthology.orchestra.run.vm06.stdout:docker (/usr/bin/docker) is present 2026-03-31T11:33:08.496 INFO:teuthology.orchestra.run.vm06.stdout:systemctl is present 2026-03-31T11:33:08.496 INFO:teuthology.orchestra.run.vm06.stdout:lvcreate is present 2026-03-31T11:33:08.498 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-31T11:33:08.499 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-31T11:33:08.501 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-31T11:33:08.501 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:08.503 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-31T11:33:08.503 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-31T11:33:08.505 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-31T11:33:08.505 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:08.507 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-31T11:33:08.507 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout masked 2026-03-31T11:33:08.508 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-31T11:33:08.508 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:08.510 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-31T11:33:08.510 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-31T11:33:08.512 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-31T11:33:08.512 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:08.515 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout enabled 2026-03-31T11:33:08.516 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout active 2026-03-31T11:33:08.516 INFO:teuthology.orchestra.run.vm06.stdout:Unit ntp.service is enabled and running 2026-03-31T11:33:08.516 INFO:teuthology.orchestra.run.vm06.stdout:Host looks OK 2026-03-31T11:33:08.517 INFO:teuthology.orchestra.run.vm06.stdout:Cluster fsid: 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:08.517 INFO:teuthology.orchestra.run.vm06.stdout:Acquiring lock 139875998821264 on /run/cephadm/48134446-2cf5-11f1-830d-4121aa8d72d5.lock 2026-03-31T11:33:08.517 INFO:teuthology.orchestra.run.vm06.stdout:Lock 139875998821264 acquired on /run/cephadm/48134446-2cf5-11f1-830d-4121aa8d72d5.lock 2026-03-31T11:33:08.517 INFO:teuthology.orchestra.run.vm06.stdout:Verifying IP 192.168.123.106 port 3300 ... 2026-03-31T11:33:08.517 INFO:teuthology.orchestra.run.vm06.stdout:Verifying IP 192.168.123.106 port 6789 ... 2026-03-31T11:33:08.517 INFO:teuthology.orchestra.run.vm06.stdout:Base mon IP(s) is [192.168.123.106:3300, 192.168.123.106:6789], mon addrv is [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-31T11:33:08.518 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.106 metric 100 2026-03-31T11:33:08.518 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-31T11:33:08.518 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout 192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.106 metric 100 2026-03-31T11:33:08.518 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout 192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.106 metric 100 2026-03-31T11:33:08.519 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-31T11:33:08.519 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout fe80::/64 dev ens3 proto kernel metric 256 pref medium 2026-03-31T11:33:08.520 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-31T11:33:08.520 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-31T11:33:08.520 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-31T11:33:08.520 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout 2: ens3: mtu 1500 state UP qlen 1000 2026-03-31T11:33:08.521 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout inet6 fe80::5055:ff:fe00:6/64 scope link 2026-03-31T11:33:08.521 INFO:teuthology.orchestra.run.vm06.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-31T11:33:08.521 INFO:teuthology.orchestra.run.vm06.stdout:Mon IP `192.168.123.106` is in CIDR network `192.168.123.0/24` 2026-03-31T11:33:08.521 INFO:teuthology.orchestra.run.vm06.stdout:Mon IP `192.168.123.106` is in CIDR network `192.168.123.0/24` 2026-03-31T11:33:08.521 INFO:teuthology.orchestra.run.vm06.stdout:Mon IP `192.168.123.106` is in CIDR network `192.168.123.1/32` 2026-03-31T11:33:08.521 INFO:teuthology.orchestra.run.vm06.stdout:Mon IP `192.168.123.106` is in CIDR network `192.168.123.1/32` 2026-03-31T11:33:08.521 INFO:teuthology.orchestra.run.vm06.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24', '192.168.123.1/32', '192.168.123.1/32'] 2026-03-31T11:33:08.521 INFO:teuthology.orchestra.run.vm06.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-31T11:33:08.522 INFO:teuthology.orchestra.run.vm06.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2... 2026-03-31T11:33:09.484 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/docker: stdout 5bb3278730741031382ca9c3dc9d221a942e06a2: Pulling from ceph-ci/ceph 2026-03-31T11:33:09.484 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/docker: stdout Digest: sha256:02c8d616f8a7af1a26efca44a51de7761356c5cb66c69c789e66f798c27c8072 2026-03-31T11:33:09.484 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/docker: stdout Status: Image is up to date for quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 2026-03-31T11:33:09.484 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/docker: stdout quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 2026-03-31T11:33:09.619 INFO:teuthology.orchestra.run.vm06.stdout:ceph: stdout ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable) 2026-03-31T11:33:09.619 INFO:teuthology.orchestra.run.vm06.stdout:Ceph version: ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable) 2026-03-31T11:33:09.619 INFO:teuthology.orchestra.run.vm06.stdout:Extracting ceph user uid/gid from container image... 2026-03-31T11:33:09.717 INFO:teuthology.orchestra.run.vm06.stdout:stat: stdout 167 167 2026-03-31T11:33:09.717 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial keys... 2026-03-31T11:33:09.806 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQD1sMtpGi2sLhAA055KyUU8gsS+e317SbboLQ== 2026-03-31T11:33:09.893 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQD1sMtpcBnqMxAAGEdaqQ9N0t2y+Vqwng4r7A== 2026-03-31T11:33:09.977 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQD1sMtpcLf9OBAAJcadDf/nTxolBf/9FqpfHA== 2026-03-31T11:33:09.977 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial monmap... 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = tentacle 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool for a [v2:192.168.123.106:3300,v1:192.168.123.106:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:setting min_mon_release = tentacle 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: set fsid to 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:10.095 INFO:teuthology.orchestra.run.vm06.stdout:Creating mon... 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.164+0000 7f2fbafc1d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.164+0000 7f2fbafc1d80 1 imported monmap: 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr epoch 0 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr last_changed 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr min_mon_release 20 (tentacle) 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr election_strategy: 1 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.164+0000 7f2fbafc1d80 0 /usr/bin/ceph-mon: set fsid to 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Git sha 0 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Compile date 2026-03-19 20:43:26 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: DB SUMMARY 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: DB Session ID: 08IQSG3US3PWQ27O9ENS 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 0, files: 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 2026-03-31T11:33:10.228 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.error_if_exists: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.create_if_missing: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.env: 0x55ccdcc1e4c0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.info_log: 0x55ccfa557d60 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.statistics: (nil) 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.use_fsync: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.db_log_dir: 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.wal_dir: 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.write_buffer_manager: 0x55ccfa54abe0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.unordered_write: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.row_cache: None 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.wal_filter: None 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.two_write_queues: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.wal_compression: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.atomic_flush: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-31T11:33:10.229 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-31T11:33:10.230 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-31T11:33:10.230 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-31T11:33:10.230 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-31T11:33:10.230 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-31T11:33:10.230 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-31T11:33:10.230 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-31T11:33:10.230 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-31T11:33:10.230 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_open_files: -1 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Compression algorithms supported: 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: kZSTD supported: 0 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: kXpressCompression supported: 0 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: kZlibCompression supported: 1 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: [db/db_impl/db_impl_open.cc:317] Creating manifest 1 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.231 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.merge_operator: 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_filter: None 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55ccfa547a80) 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks_with_high_priority: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr pin_top_level_index_and_filter: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr index_type: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr data_block_index_type: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr index_shortening: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr data_block_hash_table_util_ratio: 0.750000 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr checksum: 4 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr no_block_cache: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr block_cache: 0x55ccfa569a30 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr block_cache_name: BinnedLRUCache 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr block_cache_options: 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr capacity : 536870912 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr num_shard_bits : 4 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr strict_capacity_limit : 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr high_pri_pool_ratio: 0.000 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr block_cache_compressed: (nil) 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr persistent_cache: (nil) 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr block_size: 4096 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr block_size_deviation: 10 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr block_restart_interval: 16 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr index_block_restart_interval: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr metadata_block_size: 4096 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr partition_filters: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr use_delta_encoding: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr filter_policy: bloomfilter 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr whole_key_filtering: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr verify_compression: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr read_amp_bytes_per_bit: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr format_version: 5 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr enable_index_compression: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr block_align: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr max_auto_readahead_size: 262144 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr prepopulate_block_cache: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr initial_auto_readahead_size: 8192 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr num_file_reads_for_auto_readahead: 2 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression: NoCompression 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.num_levels: 7 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:33:10.232 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.bloom_locality: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.ttl: 2592000 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.enable_blob_files: false 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.min_blob_size: 0 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-31T11:33:10.233 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 succeeded,manifest_file_number is 1, next_file_number is 3, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 31fd3944-8735-4c92-b547-5b1b6a207f76 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.168+0000 7f2fbafc1d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 5 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.172+0000 7f2fbafc1d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55ccfa56ae00 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.172+0000 7f2fbafc1d80 4 rocksdb: DB pointer 0x55ccfa668000 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.172+0000 7f2fb274b640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.172+0000 7f2fb274b640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr ** DB Stats ** 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Flush(GB): cumulative 0.000, interval 0.000 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr AddFile(GB): cumulative 0.000, interval 0.000 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr AddFile(Total Files): cumulative 0, interval 0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr AddFile(L0 Files): cumulative 0, interval 0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr AddFile(Keys): cumulative 0, interval 0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Block cache BinnedLRUCache@0x55ccfa569a30#7 capacity: 512.00 MB usage: 0.00 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 5.7e-05 secs_since: 0 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr Block cache entry stats(count,size,portion): Misc(1,0.00 KB,0%) 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr ** File Read Latency Histogram By Level [default] ** 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.172+0000 7f2fbafc1d80 4 rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.172+0000 7f2fbafc1d80 4 rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-31T11:33:10.172+0000 7f2fbafc1d80 0 /usr/bin/ceph-mon: created monfs at /var/lib/ceph/mon/ceph-a for mon.a 2026-03-31T11:33:10.234 INFO:teuthology.orchestra.run.vm06.stdout:create mon.a on 2026-03-31T11:33:10.437 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Removed /etc/systemd/system/multi-user.target.wants/ceph.target. 2026-03-31T11:33:10.595 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-31T11:33:10.784 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5.target → /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5.target. 2026-03-31T11:33:10.784 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5.target → /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5.target. 2026-03-31T11:33:10.980 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a 2026-03-31T11:33:10.980 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to reset failed state of unit ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a.service: Unit ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a.service not loaded. 2026-03-31T11:33:11.167 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5.target.wants/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a.service → /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service. 2026-03-31T11:33:11.175 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-31T11:33:11.175 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to enable service . firewalld.service is not available 2026-03-31T11:33:11.175 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mon to start... 2026-03-31T11:33:11.175 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mon... 2026-03-31T11:33:11.537 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:11 vm06 bash[19981]: cluster 2026-03-31T11:33:11.288571+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout cluster: 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout id: 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout services: 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.235595s) [leader: a] 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout data: 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout pgs: 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:mon is available 2026-03-31T11:33:11.588 INFO:teuthology.orchestra.run.vm06.stdout:Assimilating anything we can from ceph.conf... 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [global] 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout fsid = 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = True 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [osd] 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-31T11:33:11.764 INFO:teuthology.orchestra.run.vm06.stdout:Generating new minimal ceph.conf... 2026-03-31T11:33:11.972 INFO:teuthology.orchestra.run.vm06.stdout:Restarting the monitor... 2026-03-31T11:33:12.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:11 vm06 systemd[1]: Stopping Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:33:12.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[19981]: debug 2026-03-31T11:33:12.016+0000 7f13dfd61640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-31T11:33:12.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[19981]: debug 2026-03-31T11:33:12.016+0000 7f13dfd61640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-31T11:33:12.145 INFO:teuthology.orchestra.run.vm06.stdout:Setting public_network to 192.168.123.0/24,192.168.123.1/32 in global config section 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20362]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-mon-a 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 systemd[1]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a.service: Deactivated successfully. 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 systemd[1]: Stopped Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 systemd[1]: Started Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.288+0000 7fd1a96edd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.288+0000 7fd1a96edd80 0 ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable - RelWithDebInfo), process ceph-mon, pid 7 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.288+0000 7fd1a96edd80 0 pidfile_write: ignore empty --pid-file 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 0 load: jerasure load: lrc load: isa 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Git sha 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Compile date 2026-03-19 20:43:26 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: DB SUMMARY 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: DB Session ID: FN9N4MTSJK2PI03JI9K5 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: CURRENT file: CURRENT 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 77035 ; 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.error_if_exists: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.create_if_missing: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.env: 0x5605aeee34c0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.info_log: 0x5605b81c6000 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.statistics: (nil) 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.use_fsync: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-31T11:33:12.361 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.db_log_dir: 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.wal_dir: 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.write_buffer_manager: 0x5605b81ba140 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.unordered_write: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.row_cache: None 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.wal_filter: None 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.two_write_queues: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.wal_compression: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.atomic_flush: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-31T11:33:12.362 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_open_files: -1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Compression algorithms supported: 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: kZSTD supported: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: kXpressCompression supported: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: kZlibCompression supported: 1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.merge_operator: 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_filter: None 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5605b81b6c00) 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cache_index_and_filter_blocks: 1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: pin_top_level_index_and_filter: 1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: index_type: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: data_block_index_type: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: index_shortening: 1 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: data_block_hash_table_util_ratio: 0.750000 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: checksum: 4 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: no_block_cache: 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: block_cache: 0x5605b81d9b90 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: block_cache_name: BinnedLRUCache 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: block_cache_options: 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: capacity : 536870912 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: num_shard_bits : 4 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: strict_capacity_limit : 0 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: high_pri_pool_ratio: 0.000 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: block_cache_compressed: (nil) 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: persistent_cache: (nil) 2026-03-31T11:33:12.363 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: block_size: 4096 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: block_size_deviation: 10 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: block_restart_interval: 16 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: index_block_restart_interval: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: metadata_block_size: 4096 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: partition_filters: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: use_delta_encoding: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: filter_policy: bloomfilter 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: whole_key_filtering: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: verify_compression: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: read_amp_bytes_per_bit: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: format_version: 5 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: enable_index_compression: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: block_align: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: max_auto_readahead_size: 262144 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: prepopulate_block_cache: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: initial_auto_readahead_size: 8192 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: num_file_reads_for_auto_readahead: 2 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression: NoCompression 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.num_levels: 7 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-31T11:33:12.364 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.bloom_locality: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.ttl: 2592000 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.enable_blob_files: false 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.min_blob_size: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.292+0000 7fd1a96edd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.296+0000 7fd1a96edd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.296+0000 7fd1a96edd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.296+0000 7fd1a96edd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 31fd3944-8735-4c92-b547-5b1b6a207f76 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.296+0000 7fd1a96edd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774956792297224, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.296+0000 7fd1a96edd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.296+0000 7fd1a96edd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774956792298508, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 74047, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 231, "table_properties": {"data_size": 72326, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9938, "raw_average_key_size": 49, "raw_value_size": 66671, "raw_average_value_size": 330, "num_data_blocks": 8, "num_entries": 202, "num_filter_entries": 202, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1774956792, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "31fd3944-8735-4c92-b547-5b1b6a207f76", "db_session_id": "FN9N4MTSJK2PI03JI9K5", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.296+0000 7fd1a96edd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774956792298570, "job": 1, "event": "recovery_finished"} 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.296+0000 7fd1a96edd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5605b81dae00 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 4 rocksdb: DB pointer 0x5605b832a000 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd19f45e640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd19f45e640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: ** DB Stats ** 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-31T11:33:12.365 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: ** Compaction Stats [default] ** 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: L0 2/0 74.17 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 64.4 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Sum 2/0 74.17 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 64.4 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 64.4 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: ** Compaction Stats [default] ** 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 64.4 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: AddFile(Total Files): cumulative 0, interval 0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: AddFile(Keys): cumulative 0, interval 0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Cumulative compaction: 0.00 GB write, 10.69 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Interval compaction: 0.00 GB write, 10.69 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Block cache BinnedLRUCache@0x5605b81d9b90#7 capacity: 512.00 MB usage: 1.06 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 9e-06 secs_since: 0 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: Block cache entry stats(count,size,portion): FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(1,0.00 KB,0%) 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: ** File Read Latency Histogram By Level [default] ** 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] at bind addrs [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 1 mon.a@-1(???) e1 preinit fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 0 mon.a@-1(???).mds e1 new map 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 0 mon.a@-1(???).mds e1 print_map 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: e1 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: btime 2026-03-31T11:33:11:293707+0000 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: legacy client fscid: -1 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: No filesystems configured 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 0 mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: debug 2026-03-31T11:33:12.300+0000 7fd1a96edd80 1 mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307118+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307118+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307142+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307142+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307147+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307147+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307149+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307149+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307152+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307152+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307156+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307156+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:33:12.366 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307161+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307161+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307168+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307168+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307431+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307431+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307442+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.307442+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.308132+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-31T11:33:12.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 bash[20457]: cluster 2026-03-31T11:33:12.308132+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-31T11:33:12.580 INFO:teuthology.orchestra.run.vm06.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-31T11:33:12.581 INFO:teuthology.orchestra.run.vm06.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:12.581 INFO:teuthology.orchestra.run.vm06.stdout:Creating mgr... 2026-03-31T11:33:12.581 INFO:teuthology.orchestra.run.vm06.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-31T11:33:12.581 INFO:teuthology.orchestra.run.vm06.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-31T11:33:12.786 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.a 2026-03-31T11:33:12.786 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to reset failed state of unit ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.a.service: Unit ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.a.service not loaded. 2026-03-31T11:33:12.954 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5.target.wants/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.a.service → /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service. 2026-03-31T11:33:12.957 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:12 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:12.957 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:12 vm06 systemd[1]: Started Ceph mgr.a for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:33:12.957 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:12.957 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:12 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:12.961 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-31T11:33:12.961 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to enable service . firewalld.service is not available 2026-03-31T11:33:12.961 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-31T11:33:12.961 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-31T11:33:12.961 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr to start... 2026-03-31T11:33:12.961 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr... 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "48134446-2cf5-11f1-830d-4121aa8d72d5", 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "a" 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "tentacle", 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-31T11:33:11:293707+0000", 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-31T11:33:13.176 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs" 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-31T11:33:11.294439+0000", 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-31T11:33:13.177 INFO:teuthology.orchestra.run.vm06.stdout:mgr not available, waiting (1/15)... 2026-03-31T11:33:13.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:13 vm06 bash[20457]: audit 2026-03-31T11:33:12.538815+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.106:0/4220559948' entity='client.admin' 2026-03-31T11:33:13.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:13 vm06 bash[20457]: audit 2026-03-31T11:33:12.538815+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.106:0/4220559948' entity='client.admin' 2026-03-31T11:33:13.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:13 vm06 bash[20457]: audit 2026-03-31T11:33:13.113293+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.106:0/2053872366' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-31T11:33:13.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:13 vm06 bash[20457]: audit 2026-03-31T11:33:13.113293+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.106:0/2053872366' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-31T11:33:14.815 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:14 vm06 bash[20727]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-31T11:33:14.815 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:14 vm06 bash[20727]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-31T11:33:14.816 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:14 vm06 bash[20727]: from numpy import show_config as show_numpy_config 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "48134446-2cf5-11f1-830d-4121aa8d72d5", 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "a" 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "tentacle", 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-31T11:33:15.375 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-31T11:33:11:293707+0000", 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs" 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-31T11:33:11.294439+0000", 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-31T11:33:15.376 INFO:teuthology.orchestra.run.vm06.stdout:mgr not available, waiting (2/15)... 2026-03-31T11:33:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:15 vm06 bash[20457]: audit 2026-03-31T11:33:15.338848+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.106:0/3578261567' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-31T11:33:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:15 vm06 bash[20457]: audit 2026-03-31T11:33:15.338848+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.106:0/3578261567' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-31T11:33:16.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:16 vm06 bash[20457]: cluster 2026-03-31T11:33:16.388863+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-31T11:33:16.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:16 vm06 bash[20457]: cluster 2026-03-31T11:33:16.388863+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: cluster 2026-03-31T11:33:16.393394+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00468992s) 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: cluster 2026-03-31T11:33:16.393394+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00468992s) 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.394783+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "mds metadata"} : dispatch 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.394783+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "mds metadata"} : dispatch 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.394835+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "osd metadata"} : dispatch 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.394835+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "osd metadata"} : dispatch 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.394886+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "mon metadata"} : dispatch 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.394886+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "mon metadata"} : dispatch 2026-03-31T11:33:17.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.399643+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.399643+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.399785+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "a", "id": "a"} : dispatch 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.399785+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "a", "id": "a"} : dispatch 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: cluster 2026-03-31T11:33:16.404351+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: cluster 2026-03-31T11:33:16.404351+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.412009+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"} : dispatch 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.412009+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"} : dispatch 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.413362+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"} : dispatch 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.413362+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"} : dispatch 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.415349+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.415349+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.417882+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.417882+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.420222+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' 2026-03-31T11:33:17.690 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:17 vm06 bash[20457]: audit 2026-03-31T11:33:16.420222+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.106:0/1625937629' entity='mgr.a' 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "48134446-2cf5-11f1-830d-4121aa8d72d5", 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "a" 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "tentacle", 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-31T11:33:17.721 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-31T11:33:11:293707+0000", 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs" 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-31T11:33:11.294439+0000", 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-31T11:33:17.722 INFO:teuthology.orchestra.run.vm06.stdout:mgr is available 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [global] 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout fsid = 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:18.022 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-31T11:33:18.023 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-31T11:33:18.023 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-31T11:33:18.023 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [osd] 2026-03-31T11:33:18.023 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-31T11:33:18.023 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-31T11:33:18.023 INFO:teuthology.orchestra.run.vm06.stdout:Enabling cephadm module... 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: cluster 2026-03-31T11:33:17.403348+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.01465s) 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: cluster 2026-03-31T11:33:17.403348+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.01465s) 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: audit 2026-03-31T11:33:17.678169+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.106:0/3356032322' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: audit 2026-03-31T11:33:17.678169+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.106:0/3356032322' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: audit 2026-03-31T11:33:17.972510+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.106:0/2419902254' entity='client.admin' cmd={"prefix": "config assimilate-conf"} : dispatch 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: audit 2026-03-31T11:33:17.972510+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.106:0/2419902254' entity='client.admin' cmd={"prefix": "config assimilate-conf"} : dispatch 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: audit 2026-03-31T11:33:17.975451+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.106:0/2419902254' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: audit 2026-03-31T11:33:17.975451+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.106:0/2419902254' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: audit 2026-03-31T11:33:18.318515+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.106:0/365795774' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "cephadm"} : dispatch 2026-03-31T11:33:18.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:18 vm06 bash[20457]: audit 2026-03-31T11:33:18.318515+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.106:0/365795774' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "cephadm"} : dispatch 2026-03-31T11:33:19.315 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:19 vm06 bash[20727]: ignoring --setuser ceph since I am not root 2026-03-31T11:33:19.315 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:19 vm06 bash[20727]: ignoring --setgroup ceph since I am not root 2026-03-31T11:33:19.368 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-31T11:33:19.368 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-31T11:33:19.369 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-31T11:33:19.369 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-31T11:33:19.369 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-31T11:33:19.369 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-31T11:33:19.369 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for the mgr to restart... 2026-03-31T11:33:19.369 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr epoch 4... 2026-03-31T11:33:20.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:19 vm06 bash[20457]: audit 2026-03-31T11:33:18.977316+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.106:0/365795774' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-31T11:33:20.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:19 vm06 bash[20457]: audit 2026-03-31T11:33:18.977316+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.106:0/365795774' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-31T11:33:20.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:19 vm06 bash[20457]: cluster 2026-03-31T11:33:18.980303+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-31T11:33:20.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:19 vm06 bash[20457]: cluster 2026-03-31T11:33:18.980303+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-31T11:33:20.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:19 vm06 bash[20457]: audit 2026-03-31T11:33:19.337094+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.106:0/2678475144' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-31T11:33:20.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:19 vm06 bash[20457]: audit 2026-03-31T11:33:19.337094+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.106:0/2678475144' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-31T11:33:20.816 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:20 vm06 bash[20727]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-31T11:33:20.816 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:20 vm06 bash[20727]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-31T11:33:20.816 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:20 vm06 bash[20727]: from numpy import show_config as show_numpy_config 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.475124+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.475124+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.475389+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.475389+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.479204+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.479204+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.479301+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00401209s) 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.479301+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00401209s) 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.481541+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.481541+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.481751+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "a", "id": "a"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.481751+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "a", "id": "a"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.482363+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "mds metadata"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.482363+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "mds metadata"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.482421+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "osd metadata"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.482421+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "osd metadata"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.482475+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "mon metadata"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: audit 2026-03-31T11:33:22.482475+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "mon metadata"} : dispatch 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.486608+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-31T11:33:22.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:22 vm06 bash[20457]: cluster 2026-03-31T11:33:22.486608+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-31T11:33:23.528 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-31T11:33:23.528 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-31T11:33:23.528 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-31T11:33:23.528 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-31T11:33:23.528 INFO:teuthology.orchestra.run.vm06.stdout:mgr epoch 4 is available 2026-03-31T11:33:23.528 INFO:teuthology.orchestra.run.vm06.stdout:Verifying orchestrator module is enabled... 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.161934+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.161934+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.165347+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.165347+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: cephadm 2026-03-31T11:33:23.165788+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: cephadm 2026-03-31T11:33:23.165788+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.168535+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.168535+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.174347+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.174347+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.178371+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.178371+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.184029+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.184029+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.188988+0000 mon.a (mon.0) 52 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.188988+0000 mon.a (mon.0) 52 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.193175+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.193175+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.194441+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.194441+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: cluster 2026-03-31T11:33:23.483984+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.0087s) 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: cluster 2026-03-31T11:33:23.483984+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.0087s) 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.485263+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.485263+0000 mgr.a (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.489455+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.489455+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.810148+0000 mon.a (mon.0) 56 : audit [INF] from='client.? 192.168.123.106:0/4252830660' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "orchestrator"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: audit 2026-03-31T11:33:23.810148+0000 mon.a (mon.0) 56 : audit [INF] from='client.? 192.168.123.106:0/4252830660' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "orchestrator"} : dispatch 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.015540+0000 mgr.a (mgr.14118) 4 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Bus STARTING 2026-03-31T11:33:24.499 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:24 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.015540+0000 mgr.a (mgr.14118) 4 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Bus STARTING 2026-03-31T11:33:24.565 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stderr module 'orchestrator' is already enabled (always-on) 2026-03-31T11:33:24.565 INFO:teuthology.orchestra.run.vm06.stdout:Setting orchestrator backend to cephadm... 2026-03-31T11:33:25.192 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-31T11:33:25.192 INFO:teuthology.orchestra.run.vm06.stdout:Generating ssh key... 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.125338+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Serving on https://192.168.123.106:7150 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.125338+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Serving on https://192.168.123.106:7150 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.125906+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Client ('192.168.123.106', 48430) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.125906+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Client ('192.168.123.106', 48430) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.226741+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Serving on http://192.168.123.106:8765 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.226741+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Serving on http://192.168.123.106:8765 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.226784+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Bus STARTED 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cephadm 2026-03-31T11:33:24.226784+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [31/Mar/2026:11:33:24] ENGINE Bus STARTED 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.227435+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.227435+0000 mon.a (mon.0) 57 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.484253+0000 mon.a (mon.0) 58 : audit [INF] from='client.? 192.168.123.106:0/4252830660' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "orchestrator"}]': finished 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.484253+0000 mon.a (mon.0) 58 : audit [INF] from='client.? 192.168.123.106:0/4252830660' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "orchestrator"}]': finished 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cluster 2026-03-31T11:33:24.486432+0000 mon.a (mon.0) 59 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: cluster 2026-03-31T11:33:24.486432+0000 mon.a (mon.0) 59 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.837021+0000 mgr.a (mgr.14118) 9 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.837021+0000 mgr.a (mgr.14118) 9 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.841965+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.841965+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.846065+0000 mon.a (mon.0) 61 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:25.460 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20457]: audit 2026-03-31T11:33:24.846065+0000 mon.a (mon.0) 61 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: Generating public/private ed25519 key pair. 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: Your identification has been saved in /tmp/tmpqrdyu7hc/key 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: Your public key has been saved in /tmp/tmpqrdyu7hc/key.pub 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: The key fingerprint is: 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: SHA256:eV+xRk7NKIJIUFf7e4mouYMNY8JMN/EPSMnqhvyzH0A ceph-48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: The key's randomart image is: 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: +--[ED25519 256]--+ 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | .+o.... | 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | .=o . . + | 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | Eo.+. o . = o| 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | .o + o. o = o | 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | . *.. .So. . = | 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | o *.+ ..o = . | 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | o o.= . + o | 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | o ..oo . | 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: | .+. +o | 2026-03-31T11:33:25.786 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:25 vm06 bash[20727]: +----[SHA256]-----+ 2026-03-31T11:33:25.823 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFLxinMgd9OkKgoYGGHqMXpkytxCmmT9AZSe2s17lzGD ceph-48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:25.823 INFO:teuthology.orchestra.run.vm06.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-31T11:33:25.823 INFO:teuthology.orchestra.run.vm06.stdout:Adding key to root@localhost authorized_keys... 2026-03-31T11:33:25.823 INFO:teuthology.orchestra.run.vm06.stdout:Adding host vm06... 2026-03-31T11:33:26.719 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.146709+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.146709+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.443953+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.443953+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: cephadm 2026-03-31T11:33:25.444154+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: cephadm 2026-03-31T11:33:25.444154+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.464939+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.464939+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.467852+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.467852+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.774773+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:26.720 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:26 vm06 bash[20457]: audit 2026-03-31T11:33:25.774773+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:27.753 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:27 vm06 bash[20457]: audit 2026-03-31T11:33:26.078800+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "addr": "192.168.123.106", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:27.754 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:27 vm06 bash[20457]: audit 2026-03-31T11:33:26.078800+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "addr": "192.168.123.106", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:27.754 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:27 vm06 bash[20457]: cephadm 2026-03-31T11:33:26.682412+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm06 2026-03-31T11:33:27.754 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:27 vm06 bash[20457]: cephadm 2026-03-31T11:33:26.682412+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm06 2026-03-31T11:33:28.256 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Added host 'vm06' with addr '192.168.123.106' 2026-03-31T11:33:28.256 INFO:teuthology.orchestra.run.vm06.stdout:Deploying unmanaged mon service... 2026-03-31T11:33:28.578 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-31T11:33:28.578 INFO:teuthology.orchestra.run.vm06.stdout:Deploying unmanaged mgr service... 2026-03-31T11:33:28.890 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.181927+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.181927+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: cephadm 2026-03-31T11:33:28.182484+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm06 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: cephadm 2026-03-31T11:33:28.182484+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm06 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.182889+0000 mon.a (mon.0) 65 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.182889+0000 mon.a (mon.0) 65 : audit [DBG] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.531329+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.531329+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: cephadm 2026-03-31T11:33:28.532507+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: cephadm 2026-03-31T11:33:28.532507+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.535698+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.535698+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.844467+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.844467+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: cephadm 2026-03-31T11:33:28.845230+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: cephadm 2026-03-31T11:33:28.845230+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.847751+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:28.847751+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:29.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:29.149286+0000 mon.a (mon.0) 68 : audit [INF] from='client.? 192.168.123.106:0/124728914' entity='client.admin' 2026-03-31T11:33:29.468 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:29 vm06 bash[20457]: audit 2026-03-31T11:33:29.149286+0000 mon.a (mon.0) 68 : audit [INF] from='client.? 192.168.123.106:0/124728914' entity='client.admin' 2026-03-31T11:33:29.503 INFO:teuthology.orchestra.run.vm06.stdout:Enabling the dashboard module... 2026-03-31T11:33:30.767 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20457]: audit 2026-03-31T11:33:29.453561+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.106:0/2474945648' entity='client.admin' 2026-03-31T11:33:30.767 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20457]: audit 2026-03-31T11:33:29.453561+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.106:0/2474945648' entity='client.admin' 2026-03-31T11:33:30.767 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20457]: audit 2026-03-31T11:33:29.752036+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:30.767 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20457]: audit 2026-03-31T11:33:29.752036+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:30.767 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20457]: audit 2026-03-31T11:33:29.808923+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.106:0/3460114986' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "dashboard"} : dispatch 2026-03-31T11:33:30.767 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20457]: audit 2026-03-31T11:33:29.808923+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.106:0/3460114986' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "dashboard"} : dispatch 2026-03-31T11:33:30.767 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20457]: audit 2026-03-31T11:33:30.081972+0000 mon.a (mon.0) 72 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:30.767 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20457]: audit 2026-03-31T11:33:30.081972+0000 mon.a (mon.0) 72 : audit [INF] from='mgr.14118 192.168.123.106:0/3905204634' entity='mgr.a' 2026-03-31T11:33:31.065 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20727]: ignoring --setuser ceph since I am not root 2026-03-31T11:33:31.065 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:30 vm06 bash[20727]: ignoring --setgroup ceph since I am not root 2026-03-31T11:33:31.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-31T11:33:31.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-31T11:33:31.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-31T11:33:31.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-31T11:33:31.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-31T11:33:31.177 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-31T11:33:31.177 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for the mgr to restart... 2026-03-31T11:33:31.177 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr epoch 8... 2026-03-31T11:33:32.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:31 vm06 bash[20457]: audit 2026-03-31T11:33:30.752986+0000 mon.a (mon.0) 73 : audit [INF] from='client.? 192.168.123.106:0/3460114986' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-31T11:33:32.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:31 vm06 bash[20457]: audit 2026-03-31T11:33:30.752986+0000 mon.a (mon.0) 73 : audit [INF] from='client.? 192.168.123.106:0/3460114986' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-31T11:33:32.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:31 vm06 bash[20457]: cluster 2026-03-31T11:33:30.755442+0000 mon.a (mon.0) 74 : cluster [DBG] mgrmap e8: a(active, since 8s) 2026-03-31T11:33:32.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:31 vm06 bash[20457]: cluster 2026-03-31T11:33:30.755442+0000 mon.a (mon.0) 74 : cluster [DBG] mgrmap e8: a(active, since 8s) 2026-03-31T11:33:32.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:31 vm06 bash[20457]: audit 2026-03-31T11:33:31.136517+0000 mon.a (mon.0) 75 : audit [DBG] from='client.? 192.168.123.106:0/2872804359' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-31T11:33:32.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:31 vm06 bash[20457]: audit 2026-03-31T11:33:31.136517+0000 mon.a (mon.0) 75 : audit [DBG] from='client.? 192.168.123.106:0/2872804359' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-31T11:33:32.565 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:32 vm06 bash[20727]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-31T11:33:32.565 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:32 vm06 bash[20727]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-31T11:33:32.565 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:32 vm06 bash[20727]: from numpy import show_config as show_numpy_config 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.349012+0000 mon.a (mon.0) 76 : cluster [INF] Active manager daemon a restarted 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.349012+0000 mon.a (mon.0) 76 : cluster [INF] Active manager daemon a restarted 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.349294+0000 mon.a (mon.0) 77 : cluster [INF] Activating manager daemon a 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.349294+0000 mon.a (mon.0) 77 : cluster [INF] Activating manager daemon a 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.354175+0000 mon.a (mon.0) 78 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.354175+0000 mon.a (mon.0) 78 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.354228+0000 mon.a (mon.0) 79 : cluster [DBG] mgrmap e9: a(active, starting, since 0.0050738s) 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.354228+0000 mon.a (mon.0) 79 : cluster [DBG] mgrmap e9: a(active, starting, since 0.0050738s) 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.354804+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.354804+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.355674+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "a", "id": "a"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.355674+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "a", "id": "a"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.356555+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mds metadata"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.356555+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mds metadata"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.356677+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.356677+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.356798+0000 mon.a (mon.0) 84 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: audit 2026-03-31T11:33:34.356798+0000 mon.a (mon.0) 84 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata"} : dispatch 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.363127+0000 mon.a (mon.0) 85 : cluster [INF] Manager daemon a is now available 2026-03-31T11:33:34.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:34 vm06 bash[20457]: cluster 2026-03-31T11:33:34.363127+0000 mon.a (mon.0) 85 : cluster [INF] Manager daemon a is now available 2026-03-31T11:33:35.406 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-31T11:33:35.406 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 9, 2026-03-31T11:33:35.406 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-31T11:33:35.406 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-31T11:33:35.406 INFO:teuthology.orchestra.run.vm06.stdout:mgr epoch 8 is available 2026-03-31T11:33:35.406 INFO:teuthology.orchestra.run.vm06.stdout:Using certmgr to generate dashboard self-signed certificate... 2026-03-31T11:33:35.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: audit 2026-03-31T11:33:34.569394+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:35.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: audit 2026-03-31T11:33:34.569394+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:35.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: audit 2026-03-31T11:33:34.571111+0000 mon.a (mon.0) 87 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:35.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: audit 2026-03-31T11:33:34.571111+0000 mon.a (mon.0) 87 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:35.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: audit 2026-03-31T11:33:34.593886+0000 mon.a (mon.0) 88 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"} : dispatch 2026-03-31T11:33:35.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: audit 2026-03-31T11:33:34.593886+0000 mon.a (mon.0) 88 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"} : dispatch 2026-03-31T11:33:35.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: audit 2026-03-31T11:33:34.598049+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"} : dispatch 2026-03-31T11:33:35.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: audit 2026-03-31T11:33:34.598049+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"} : dispatch 2026-03-31T11:33:35.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: cluster 2026-03-31T11:33:35.363402+0000 mon.a (mon.0) 90 : cluster [DBG] mgrmap e10: a(active, since 1.01425s) 2026-03-31T11:33:35.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:35 vm06 bash[20457]: cluster 2026-03-31T11:33:35.363402+0000 mon.a (mon.0) 90 : cluster [DBG] mgrmap e10: a(active, since 1.01425s) 2026-03-31T11:33:36.161 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout {"cert": "-----BEGIN CERTIFICATE-----\nMIIE+TCCAuGgAwIBAgIUfa2DzNNFeiHHevxP6fRTAIRrROcwDQYJKoZIhvcNAQEL\nBQAwFzEVMBMGA1UEAwwMY2VwaGFkbS1yb290MB4XDTI2MDMzMTExMzMzNloXDTI5\nMDMzMDExMzMzNlowGjEYMBYGA1UEAwwPMTkyLjE2OC4xMjMuMTA2MIICIjANBgkq\nhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEApYybgcI6ppDUjTzbGzvp5Rr56U2JV2zv\ncfdR9FmvBwJSsXZsbIXz3PTNpkxvJg5F+f23fH5nTs+4bGYb8ss7P1396EbJyfaT\nHEXH/FvjBEssJ/zU1WQJltdQVrmHQTVwGafEuzCavn+hIPTNbjUbvZAZJ7RTTvfk\nwCwTjHH/CiS9w2EmbebdQyHinp/0q05D/xZmMazbKuuLcw5bYVxbG/ac0ClQ7eX7\na3DhiVHMEK81FB1haLQ28vqMWqhNXClmoCDe7B8yW8rQK329LLWgVr6QLo5XA6oC\nKmzAwlpSWwfzcVsxr8equAeIfujsETpL7gdqwkVg2/Bp/PP9d+AUSrE9RFjhBCiM\nmK1mVDr04kZfaMqJxhHo7FF0C1Pp8YPdhm6OsSzVNgzUzCQiwrrO2bcpPPwLiaaK\n63OCM0KG0bRAQB9/M6WF7fW9u7x/wBwM9mDPf/mi13KcAORz8dNAzb1Gu6kJ0yMo\nN5WtmYgOlWThM4ZPiFuKSIVmnoWu+tjL68Gt4bqkalsyfiAUgqMiuzQzGGfZ/Xev\noQHD5o6slnLPT0HWbVJTtcddRsGTjQ5yUTEN63dY0rgT5/7LQi4fw7gMRnSO2Adr\nU+mR5Gbln8y/XNvoHm0h3fcafgtEdZ2y6redZebX7MObuISOIXXjTHY4XF8mrOtc\ntQd0+nN0oDsCAwEAAaM6MDgwKAYDVR0RBCEwH4IEdm0wNoIRZGFzaGJvYXJkX3Nl\ncnZlcnOHBMCoe2owDAYDVR0TAQH/BAIwADANBgkqhkiG9w0BAQsFAAOCAgEAAYX9\nXuKUhiPl5RbjPYtKfJTQEOG66c5wKnJOysPHoULvL0CdQ6XSwAvkIczvJMcKiQr/\nSwp63dnWdHtrmXHn0OK46To1ryUt1pzTKlxPCntq/MCpI+wwnimES6Kmmgc39+Cz\nRt5elG3NzPluPZrpN5PmwLzD+Ssh4iGr4Ot0JiR39yp+x+lqpmSyBLkQm9wpPI5m\n8AwWmwkXTjokkIHFToEMiNRncWZGp1Yb/Nzc3s2d2ZZGlaRTsNRsAjsGgkclB9+q\nkoBHJP9knTnges1xVKTvWFk04ImjnFGW6I+FzKgXqUkbWhU8iKVy744RWoFBamHc\ncu845Hn5d0UYzlQtl34oUjt6G8w4nMAk1gsvuSP65yYmti855/fhDAIWzPq7SB5I\nMQYUSHhg7hekSln6713PfxnLDNbAP7+4UJLLYX/Qgl0tjCtWBfR/0AP7q738c113\naLR1PyCAit1iPtTib4z6V7+y7yLVPqMWIanhFfuOtxo2yDer24tse8AaxzhGywKJ\nJOk6IEDizfQgAQH+UccgzL4CMWiqztUgntnHys8zllXrJYmblZHE+yhVqjiVA86u\nz8dX+LbJY9AtC2O2tE01bShMPPTz4UNcC8vYNUWixRreMYYK66+U+MSOWbomVrC7\nuxG6iuF3rL/M9NpRUHHPih/VqlTyhY0anjxJH5Q=\n-----END CERTIFICATE-----\n", "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIJKQIBAAKCAgEApYybgcI6ppDUjTzbGzvp5Rr56U2JV2zvcfdR9FmvBwJSsXZs\nbIXz3PTNpkxvJg5F+f23fH5nTs+4bGYb8ss7P1396EbJyfaTHEXH/FvjBEssJ/zU\n1WQJltdQVrmHQTVwGafEuzCavn+hIPTNbjUbvZAZJ7RTTvfkwCwTjHH/CiS9w2Em\nbebdQyHinp/0q05D/xZmMazbKuuLcw5bYVxbG/ac0ClQ7eX7a3DhiVHMEK81FB1h\naLQ28vqMWqhNXClmoCDe7B8yW8rQK329LLWgVr6QLo5XA6oCKmzAwlpSWwfzcVsx\nr8equAeIfujsETpL7gdqwkVg2/Bp/PP9d+AUSrE9RFjhBCiMmK1mVDr04kZfaMqJ\nxhHo7FF0C1Pp8YPdhm6OsSzVNgzUzCQiwrrO2bcpPPwLiaaK63OCM0KG0bRAQB9/\nM6WF7fW9u7x/wBwM9mDPf/mi13KcAORz8dNAzb1Gu6kJ0yMoN5WtmYgOlWThM4ZP\niFuKSIVmnoWu+tjL68Gt4bqkalsyfiAUgqMiuzQzGGfZ/XevoQHD5o6slnLPT0HW\nbVJTtcddRsGTjQ5yUTEN63dY0rgT5/7LQi4fw7gMRnSO2AdrU+mR5Gbln8y/XNvo\nHm0h3fcafgtEdZ2y6redZebX7MObuISOIXXjTHY4XF8mrOtctQd0+nN0oDsCAwEA\nAQKCAgAAzMcg6gwCgLH7j2RiTfh5WkIENehd+xwaaYmEgWfDxD0bYmkaX839Lqmi\nY2gov5ToT6jdTyG3Uxr0WyHzcGFJkxiw+0vqT+iyEEbrxJM6PWq/NAzgjb/bchA5\n+wzY7HsthI6sy4qpNPtVqhkmroGtzs5wT1dqYERDXB3UEJDyPu8zzHpfuZKEH9vK\ndxdljyvrmvNvdgMGbWZFGLf++8VCQ1TKpLMhaHpohdPGfWtbq/JjtdNMquJBFk4r\nT041uHV90erXKpIXDh8nYU99XHtOhVPeWw23P3u3MhiprRS3dThglII5notdTEav\n5kw4YRcQ1V8CBU0h2ELgLVr3g1EAcWWYyxccc7Nmuy3NO2M39iCUp45Q33OzINQT\nfMag8i1W8xxgatrJV+nJQbcESaKgOHM6kkvvazeFlT4o5XipW3ef6ji3oyfrbhOl\nQuBuKgRUpWe0jZaUNErLk7yz93f0uD+FuFqhCmOj0rWr+vxUhiijmXFaRBUeehKY\nxjahKj7kbTQVP61L+i93l7y5I7jPqTfG55uSCZAfa1nSGSBPzMXIPzynI8OXomX0\nRkWRd8OkPTSLkCW8YCOf7/gJC0udGwIGAKxT23+fTTKxdokLkRfQJ1qW9Xy4w1HT\nadknqUA5cZZmU+JGxg7r9L7Gw0Q0ttTIizad5e1GwW2Z8hPegQKCAQEA3LFD7P3E\nLVus+/nOFgBu9R+Z7pTeLM4/ctKPx+hh2aSEDEpZ6E8Cokvp0PPd6ervW3cWLlwg\nlyPa4Wib8XW+6H4tlILOLs463RMtG1pYVPxPW7Xq+4wNWqbXsdHhl4zi+3R6vN7r\nVE8R/DlSHsLG2vLSCTjXNm3e3VkRvq/AswFNiFRAdl2ep7kY2FkWk2bmOIr0i0H5\ngw5N0TbkMFmNXTe6ckyIn/zLKnTh2b1I4dQ4dDi76tybgIUqRVT7UoFMYfPGeHv7\ntBXorM/uUJ9nrq717q+F/0Lia0ZTTVuvnZMk0FLrwl+pCiDECXLsAgJlNlgM5udi\nm7GgDNGdgVNZ2wKCAQEAwAjiOU3K6YZDnINd1vtxsL5eG9iIsF0JJEJfXy36zY/y\nY+O2NHXfUEp/j9z4xNV7AQ8aQJw9A1MEGIF+0sV9Z0JNaZlYF66aTsT2OWDtEzIR\n8fPjtghn1wPRdTxxGuO7Ki5wcyk9N2SQB0TVKXMB4erNwDn1gk6fIHK57tJ1Q2Mi\nukkYkXGm7qsoe8EuQvk7XAs1EQbJXzZbsw0vadmQBcLzYWeim6agyOUn0mf+HKFT\n4E7n2QsnJk+fVT1oRlZ2S3tjZXC/XOqZdm3j2PZfWU+HZeIcaL1dmeeD7n51XohO\naMPPER5EiCS2O3USYu1QlW+jAlt+sESrUJTHYMORIQKCAQEAl6csgS+E3pJv1Zs8\nlfvoQ4NauQsNfYahLRsqeHnku7WElBmHnM6mhbZazx61cOfjYVv9HxBpyNJ3t4u2\nFYAPFx6V3cbhPmKzKnXh+jjkGEplkiT5e7ArAQ9Di0splYXX+Gl7akxZNE89kbAm\nEUNReOWBlZq8sQdhLYPmYL3PUnLjv3n97hi5TJIyU4EstU/xhgterk9A0WFD8i+h\nMORo8xS3h5G2FRKW3e8BwCmFasH1pU/TOjoslW1stcYHChQV0/27ddHlDkQ75nzg\nTACuM+jA0StGjjl887CYz8DgdgzGMG262o/3ziDj9YZGTb8to+FvxEBej5TI4eYt\nMNCBHwKCAQBvGDSX5jNPagdPUohFSO36zAbxoRFixk/1lBhMnfMMXQxjUWclblOk\nxZSKi9G3PbGNaUlMu5Zoh+xEqRXPpcToqejTAI2AFdjbXQL9Kt+8TSkoimnf83jv\nYRwZ8L+WSM0MmBUh6mep1hmkuuN+uTQLmKUzYdRwcD+KrBR433RLady7XHJEIjXq\nuUYQYi/7OuhmVW17zitNU8s1CJm3aM/zJVdbsHLYDUvTLQ8eoF7X/VTi3YMvg05A\n0athysa6emaxQE6+br2maWoio2w/Jjj/GA1Cmc7JoVO3J8KhhLMt3zlNTopKYvJO\nuL5cI1xhstz9trndMUVXgEKDua6mgj1BAoIBAQCNm4cknKV/yqGfJi4RdK1AQurS\nO2SGULulEF9g1CU8dpGsJCUq10sg3urgFE6UdkiFhg0+4nccOTbO3cO+V0VC15QU\n5IBcGc/dqgBCbvZGK7DbkCm9JGZ/fQMy2HbxABqv2/pKe9gd4iGCO/fZcWgQDc9m\nDpRRCiEpMxt1lPQbJ29XcYUOKbJqknxKIWs/wqx3OXN8K1dg4fkCt+70AcFmsNIN\nIEiZm2u16UGWmQlzCq49AwKDNBz8bPZ4pxXclGKRjbg1NC/AEJiRoTyvrSTCbi4N\nKV8Y/I3hTp6RvODCQsze8g3oof99VIUcXSJoa4VaLqgPpj3uBI/0qqibcqjn\n-----END RSA PRIVATE KEY-----\n"} 2026-03-31T11:33:36.451 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout SSL certificate updated 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.406498+0000 mgr.a (mgr.14152) 3 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Bus STARTING 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.406498+0000 mgr.a (mgr.14152) 3 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Bus STARTING 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.508271+0000 mgr.a (mgr.14152) 4 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Serving on http://192.168.123.106:8765 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.508271+0000 mgr.a (mgr.14152) 4 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Serving on http://192.168.123.106:8765 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.616942+0000 mgr.a (mgr.14152) 5 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Serving on https://192.168.123.106:7150 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.616942+0000 mgr.a (mgr.14152) 5 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Serving on https://192.168.123.106:7150 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.617097+0000 mgr.a (mgr.14152) 6 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Bus STARTED 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.617097+0000 mgr.a (mgr.14152) 6 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Bus STARTED 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.617324+0000 mgr.a (mgr.14152) 7 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Client ('192.168.123.106', 34882) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: cephadm 2026-03-31T11:33:35.617324+0000 mgr.a (mgr.14152) 7 : cephadm [INF] [31/Mar/2026:11:33:35] ENGINE Client ('192.168.123.106', 34882) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-31T11:33:36.731 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: audit 2026-03-31T11:33:35.672028+0000 mgr.a (mgr.14152) 8 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "orch certmgr generate-certificates", "module_name": "dashboard", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:36.732 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: audit 2026-03-31T11:33:35.672028+0000 mgr.a (mgr.14152) 8 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "orch certmgr generate-certificates", "module_name": "dashboard", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:36.732 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: audit 2026-03-31T11:33:36.409219+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:36.732 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:36 vm06 bash[20457]: audit 2026-03-31T11:33:36.409219+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:36.780 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout SSL certificate key updated 2026-03-31T11:33:36.780 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial admin user... 2026-03-31T11:33:37.339 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$X04NNVLdrUFgd1RjCFUKN.RJccojmJ6sO89MGTJMdDRdsbI93AI8i", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1774956817, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-31T11:33:37.339 INFO:teuthology.orchestra.run.vm06.stdout:Fetching dashboard port number... 2026-03-31T11:33:37.628 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 8443 2026-03-31T11:33:37.628 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-31T11:33:37.628 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-31T11:33:37.629 INFO:teuthology.orchestra.run.vm06.stdout:Ceph Dashboard is now available at: 2026-03-31T11:33:37.629 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:37.629 INFO:teuthology.orchestra.run.vm06.stdout: URL: https://vm06.local:8443/ 2026-03-31T11:33:37.629 INFO:teuthology.orchestra.run.vm06.stdout: User: admin 2026-03-31T11:33:37.629 INFO:teuthology.orchestra.run.vm06.stdout: Password: hzq6lc3l44 2026-03-31T11:33:37.629 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:37.629 INFO:teuthology.orchestra.run.vm06.stdout:Saving cluster configuration to /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config directory 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: audit 2026-03-31T11:33:36.406285+0000 mgr.a (mgr.14152) 9 : audit [DBG] from='client.14166 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: audit 2026-03-31T11:33:36.406285+0000 mgr.a (mgr.14152) 9 : audit [DBG] from='client.14166 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: cluster 2026-03-31T11:33:36.580176+0000 mon.a (mon.0) 92 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: cluster 2026-03-31T11:33:36.580176+0000 mon.a (mon.0) 92 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: audit 2026-03-31T11:33:36.716647+0000 mgr.a (mgr.14152) 10 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: audit 2026-03-31T11:33:36.716647+0000 mgr.a (mgr.14152) 10 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: audit 2026-03-31T11:33:36.720005+0000 mon.a (mon.0) 93 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: audit 2026-03-31T11:33:36.720005+0000 mon.a (mon.0) 93 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: audit 2026-03-31T11:33:37.297512+0000 mon.a (mon.0) 94 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:37.983 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:37 vm06 bash[20457]: audit 2026-03-31T11:33:37.297512+0000 mon.a (mon.0) 94 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout:Or, if you are only running a single cluster on this host: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: ceph telemetry on 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout:For more information see: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:38.012 INFO:teuthology.orchestra.run.vm06.stdout:Bootstrap complete. 2026-03-31T11:33:38.018 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout static 2026-03-31T11:33:38.021 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active logrotate 2026-03-31T11:33:38.021 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-31T11:33:38.021 INFO:teuthology.orchestra.run.vm06.stdout:Enabling the logrotate.timer service to perform daily log rotation. 2026-03-31T11:33:38.234 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:38 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:38.234 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:38 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:38.240 INFO:tasks.cephadm:Fetching config... 2026-03-31T11:33:38.240 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:33:38.240 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-31T11:33:38.243 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-31T11:33:38.243 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:33:38.243 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-31T11:33:38.292 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-31T11:33:38.292 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:33:38.292 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/keyring of=/dev/stdout 2026-03-31T11:33:38.344 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-31T11:33:38.344 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:33:38.344 DEBUG:teuthology.orchestra.run.vm06:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-31T11:33:38.393 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-31T11:33:38.393 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFLxinMgd9OkKgoYGGHqMXpkytxCmmT9AZSe2s17lzGD ceph-48134446-2cf5-11f1-830d-4121aa8d72d5' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-31T11:33:38.449 INFO:teuthology.orchestra.run.vm06.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFLxinMgd9OkKgoYGGHqMXpkytxCmmT9AZSe2s17lzGD ceph-48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:38.453 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFLxinMgd9OkKgoYGGHqMXpkytxCmmT9AZSe2s17lzGD ceph-48134446-2cf5-11f1-830d-4121aa8d72d5' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-31T11:33:38.465 INFO:teuthology.orchestra.run.vm09.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFLxinMgd9OkKgoYGGHqMXpkytxCmmT9AZSe2s17lzGD ceph-48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:38.471 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-31T11:33:38.778 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:38.796 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:38 vm06 bash[20457]: audit 2026-03-31T11:33:37.052155+0000 mgr.a (mgr.14152) 11 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:38.796 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:38 vm06 bash[20457]: audit 2026-03-31T11:33:37.052155+0000 mgr.a (mgr.14152) 11 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:38.796 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:38 vm06 bash[20457]: audit 2026-03-31T11:33:37.582129+0000 mon.a (mon.0) 95 : audit [DBG] from='client.? 192.168.123.106:0/1489049707' entity='client.admin' cmd={"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"} : dispatch 2026-03-31T11:33:38.796 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:38 vm06 bash[20457]: audit 2026-03-31T11:33:37.582129+0000 mon.a (mon.0) 95 : audit [DBG] from='client.? 192.168.123.106:0/1489049707' entity='client.admin' cmd={"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"} : dispatch 2026-03-31T11:33:38.796 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:38 vm06 bash[20457]: audit 2026-03-31T11:33:37.971560+0000 mon.a (mon.0) 96 : audit [INF] from='client.? 192.168.123.106:0/3109615328' entity='client.admin' 2026-03-31T11:33:38.796 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:38 vm06 bash[20457]: audit 2026-03-31T11:33:37.971560+0000 mon.a (mon.0) 96 : audit [INF] from='client.? 192.168.123.106:0/3109615328' entity='client.admin' 2026-03-31T11:33:39.198 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-31T11:33:39.198 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-31T11:33:39.448 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:39.898 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm09 2026-03-31T11:33:39.898 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:33:39.898 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.conf 2026-03-31T11:33:39.902 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:33:39.902 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:39.949 INFO:tasks.cephadm:Adding host vm09 to orchestrator... 2026-03-31T11:33:39.949 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch host add vm09 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:39 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:39 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.049347+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.049347+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.049951+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"} : dispatch 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.049951+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"} : dispatch 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.053111+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.053111+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.054261+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "client.agent.vm06", "caps": []} : dispatch 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.054261+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "client.agent.vm06", "caps": []} : dispatch 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.055433+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm06", "caps": []}]': finished 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.055433+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm06", "caps": []}]': finished 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.108981+0000 mon.a (mon.0) 102 : audit [INF] from='client.? 192.168.123.106:0/3875436024' entity='client.admin' 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.108981+0000 mon.a (mon.0) 102 : audit [INF] from='client.? 192.168.123.106:0/3875436024' entity='client.admin' 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: cephadm 2026-03-31T11:33:39.533880+0000 mgr.a (mgr.14152) 12 : cephadm [INF] Deploying daemon agent.vm06 on vm06 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: cephadm 2026-03-31T11:33:39.533880+0000 mgr.a (mgr.14152) 12 : cephadm [INF] Deploying daemon agent.vm06 on vm06 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.794337+0000 mgr.a (mgr.14152) 13 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.794337+0000 mgr.a (mgr.14152) 13 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.798068+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:40.163 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 bash[20457]: audit 2026-03-31T11:33:39.798068+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:40.227 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:40.539 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:40 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:40.539 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:40 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.290053+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.290053+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.296355+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.296355+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.299427+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.299427+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.305562+0000 mon.a (mon.0) 107 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.305562+0000 mon.a (mon.0) 107 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.307772+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.307772+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.309039+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.309039+0000 mon.a (mon.0) 109 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: cephadm 2026-03-31T11:33:40.309982+0000 mgr.a (mgr.14152) 14 : cephadm [INF] Updating vm06:/etc/ceph/ceph.conf 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: cephadm 2026-03-31T11:33:40.309982+0000 mgr.a (mgr.14152) 14 : cephadm [INF] Updating vm06:/etc/ceph/ceph.conf 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: cephadm 2026-03-31T11:33:40.347683+0000 mgr.a (mgr.14152) 15 : cephadm [INF] Updating vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: cephadm 2026-03-31T11:33:40.347683+0000 mgr.a (mgr.14152) 15 : cephadm [INF] Updating vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: cephadm 2026-03-31T11:33:40.382982+0000 mgr.a (mgr.14152) 16 : cephadm [INF] Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: cephadm 2026-03-31T11:33:40.382982+0000 mgr.a (mgr.14152) 16 : cephadm [INF] Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: cephadm 2026-03-31T11:33:40.421681+0000 mgr.a (mgr.14152) 17 : cephadm [INF] Updating vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.client.admin.keyring 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: cephadm 2026-03-31T11:33:40.421681+0000 mgr.a (mgr.14152) 17 : cephadm [INF] Updating vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.client.admin.keyring 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.460507+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.460507+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.464913+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.464913+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.472520+0000 mon.a (mon.0) 112 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.472520+0000 mon.a (mon.0) 112 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.473857+0000 mon.a (mon.0) 113 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.473857+0000 mon.a (mon.0) 113 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.474307+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:41.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.474307+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:41.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.477150+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.477150+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.598210+0000 mgr.a (mgr.14152) 18 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:41.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.598210+0000 mgr.a (mgr.14152) 18 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:41.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.645239+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:41.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:41 vm06 bash[20457]: audit 2026-03-31T11:33:40.645239+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:42.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:42 vm06 bash[20457]: cephadm 2026-03-31T11:33:41.194122+0000 mgr.a (mgr.14152) 19 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-31T11:33:42.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:42 vm06 bash[20457]: cephadm 2026-03-31T11:33:41.194122+0000 mgr.a (mgr.14152) 19 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-31T11:33:42.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:42 vm06 bash[20457]: cluster 2026-03-31T11:33:41.305292+0000 mon.a (mon.0) 117 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-31T11:33:42.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:42 vm06 bash[20457]: cluster 2026-03-31T11:33:41.305292+0000 mon.a (mon.0) 117 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-31T11:33:42.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:42 vm06 bash[20457]: audit 2026-03-31T11:33:41.411328+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:42.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:42 vm06 bash[20457]: audit 2026-03-31T11:33:41.411328+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:42.711 INFO:teuthology.orchestra.run.vm06.stdout:Added host 'vm09' with addr '192.168.123.109' 2026-03-31T11:33:42.814 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch host ls --format=json 2026-03-31T11:33:43.116 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:43.485 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:43.486 INFO:teuthology.orchestra.run.vm06.stdout:[{"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}, {"addr": "192.168.123.109", "hostname": "vm09", "labels": [], "status": ""}] 2026-03-31T11:33:43.564 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-31T11:33:43.564 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd crush tunables default 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:42.708326+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:42.708326+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: cephadm 2026-03-31T11:33:42.708802+0000 mgr.a (mgr.14152) 20 : cephadm [INF] Added host vm09 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: cephadm 2026-03-31T11:33:42.708802+0000 mgr.a (mgr.14152) 20 : cephadm [INF] Added host vm09 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:42.709036+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:42.709036+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.062116+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.062116+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.377941+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.377941+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.380718+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.380718+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.381249+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.381249+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.381987+0000 mon.a (mon.0) 125 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.381987+0000 mon.a (mon.0) 125 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.382524+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.382524+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.556192+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.556192+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.559222+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.559222+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.561745+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.561745+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []} : dispatch 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.563486+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-31T11:33:43.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:43 vm06 bash[20457]: audit 2026-03-31T11:33:43.563486+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-31T11:33:43.851 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:44.715 INFO:teuthology.orchestra.run.vm06.stderr:adjusted tunables profile to default 2026-03-31T11:33:44.811 INFO:tasks.cephadm:Adding mon.a on vm06 2026-03-31T11:33:44.811 INFO:tasks.cephadm:Adding mon.b on vm09 2026-03-31T11:33:44.811 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch apply mon '2;vm06:192.168.123.106=a;vm09:192.168.123.109=b' 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: cephadm 2026-03-31T11:33:43.383395+0000 mgr.a (mgr.14152) 21 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: cephadm 2026-03-31T11:33:43.383395+0000 mgr.a (mgr.14152) 21 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: cephadm 2026-03-31T11:33:43.424503+0000 mgr.a (mgr.14152) 22 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: cephadm 2026-03-31T11:33:43.424503+0000 mgr.a (mgr.14152) 22 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: cephadm 2026-03-31T11:33:43.466418+0000 mgr.a (mgr.14152) 23 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: cephadm 2026-03-31T11:33:43.466418+0000 mgr.a (mgr.14152) 23 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: audit 2026-03-31T11:33:43.485326+0000 mgr.a (mgr.14152) 24 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: audit 2026-03-31T11:33:43.485326+0000 mgr.a (mgr.14152) 24 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: cephadm 2026-03-31T11:33:43.506860+0000 mgr.a (mgr.14152) 25 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.client.admin.keyring 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: cephadm 2026-03-31T11:33:43.506860+0000 mgr.a (mgr.14152) 25 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.client.admin.keyring 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: audit 2026-03-31T11:33:44.183418+0000 mon.a (mon.0) 131 : audit [INF] from='client.? 192.168.123.106:0/4278457068' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: audit 2026-03-31T11:33:44.183418+0000 mon.a (mon.0) 131 : audit [INF] from='client.? 192.168.123.106:0/4278457068' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: audit 2026-03-31T11:33:44.591069+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:45.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:44 vm06 bash[20457]: audit 2026-03-31T11:33:44.591069+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:45.101 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:45.512 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mon update... 2026-03-31T11:33:45.618 DEBUG:teuthology.orchestra.run.vm09:mon.b> sudo journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.b.service 2026-03-31T11:33:45.619 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-31T11:33:45.619 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph mon dump -f json 2026-03-31T11:33:45.956 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: cephadm 2026-03-31T11:33:44.009417+0000 mgr.a (mgr.14152) 26 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: cephadm 2026-03-31T11:33:44.009417+0000 mgr.a (mgr.14152) 26 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.714972+0000 mon.a (mon.0) 133 : audit [INF] from='client.? 192.168.123.106:0/4278457068' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.714972+0000 mon.a (mon.0) 133 : audit [INF] from='client.? 192.168.123.106:0/4278457068' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: cluster 2026-03-31T11:33:44.718097+0000 mon.a (mon.0) 134 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: cluster 2026-03-31T11:33:44.718097+0000 mon.a (mon.0) 134 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.739421+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.739421+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.742544+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.742544+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.747964+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.747964+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.764004+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.764004+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.766569+0000 mon.a (mon.0) 139 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.766569+0000 mon.a (mon.0) 139 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.767150+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.767150+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.772350+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.772350+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.786078+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.786078+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.787519+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.787519+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.787949+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.787949+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.791954+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:44.791954+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.065483+0000 mon.a (mon.0) 146 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.065483+0000 mon.a (mon.0) 146 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.143640+0000 mon.a (mon.0) 147 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.143640+0000 mon.a (mon.0) 147 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.147136+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.147136+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.403759+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.403759+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.407726+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.407726+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.411162+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.411162+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.511722+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.511722+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.512829+0000 mon.a (mon.0) 153 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.512829+0000 mon.a (mon.0) 153 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.514250+0000 mon.a (mon.0) 154 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.514250+0000 mon.a (mon.0) 154 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.514702+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.514702+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.518256+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.518256+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.519717+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.519717+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.520333+0000 mon.a (mon.0) 158 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:46.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:45 vm06 bash[20457]: audit 2026-03-31T11:33:45.520333+0000 mon.a (mon.0) 158 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:46.370 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:33:46.370 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"48134446-2cf5-11f1-830d-4121aa8d72d5","modified":"2026-03-31T11:33:10.045398Z","created":"2026-03-31T11:33:10.045398Z","min_mon_release":20,"min_mon_release_name":"tentacle","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid","tentacle"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-31T11:33:46.370 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: audit 2026-03-31T11:33:45.505345+0000 mgr.a (mgr.14152) 27 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm06:192.168.123.106=a;vm09:192.168.123.109=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: audit 2026-03-31T11:33:45.505345+0000 mgr.a (mgr.14152) 27 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm06:192.168.123.106=a;vm09:192.168.123.109=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: cephadm 2026-03-31T11:33:45.507270+0000 mgr.a (mgr.14152) 28 : cephadm [INF] Saving service mon spec with placement vm06:192.168.123.106=a;vm09:192.168.123.109=b;count:2 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: cephadm 2026-03-31T11:33:45.507270+0000 mgr.a (mgr.14152) 28 : cephadm [INF] Saving service mon spec with placement vm06:192.168.123.106=a;vm09:192.168.123.109=b;count:2 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: cephadm 2026-03-31T11:33:45.521191+0000 mgr.a (mgr.14152) 29 : cephadm [INF] Deploying daemon mon.b on vm09 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: cephadm 2026-03-31T11:33:45.521191+0000 mgr.a (mgr.14152) 29 : cephadm [INF] Deploying daemon mon.b on vm09 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: audit 2026-03-31T11:33:45.859631+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: audit 2026-03-31T11:33:45.859631+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: audit 2026-03-31T11:33:45.865985+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: audit 2026-03-31T11:33:45.865985+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: audit 2026-03-31T11:33:46.370021+0000 mon.a (mon.0) 161 : audit [DBG] from='client.? 192.168.123.109:0/2703158535' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-31T11:33:47.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:46 vm06 bash[20457]: audit 2026-03-31T11:33:46.370021+0000 mon.a (mon.0) 161 : audit [DBG] from='client.? 192.168.123.109:0/2703158535' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-31T11:33:47.308 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.380718+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.308 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.380718+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.564 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.381249+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"} : dispatch 2026-03-31T11:33:47.564 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.381249+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"} : dispatch 2026-03-31T11:33:47.564 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.381987+0000 mon.a (mon.0) 125 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.564 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.381987+0000 mon.a (mon.0) 125 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.564 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.382524+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.382524+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.556192+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.556192+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.559222+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.559222+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.561745+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.561745+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.563486+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.563486+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:43.383395+0000 mgr.a (mgr.14152) 21 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:43.383395+0000 mgr.a (mgr.14152) 21 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:43.424503+0000 mgr.a (mgr.14152) 22 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:43.424503+0000 mgr.a (mgr.14152) 22 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:43.466418+0000 mgr.a (mgr.14152) 23 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:43.466418+0000 mgr.a (mgr.14152) 23 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.485326+0000 mgr.a (mgr.14152) 24 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:43.485326+0000 mgr.a (mgr.14152) 24 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:43.506860+0000 mgr.a (mgr.14152) 25 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.client.admin.keyring 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:43.506860+0000 mgr.a (mgr.14152) 25 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.client.admin.keyring 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.183418+0000 mon.a (mon.0) 131 : audit [INF] from='client.? 192.168.123.106:0/4278457068' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.183418+0000 mon.a (mon.0) 131 : audit [INF] from='client.? 192.168.123.106:0/4278457068' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.591069+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.591069+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:44.009417+0000 mgr.a (mgr.14152) 26 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:44.009417+0000 mgr.a (mgr.14152) 26 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.714972+0000 mon.a (mon.0) 133 : audit [INF] from='client.? 192.168.123.106:0/4278457068' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.714972+0000 mon.a (mon.0) 133 : audit [INF] from='client.? 192.168.123.106:0/4278457068' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cluster 2026-03-31T11:33:44.718097+0000 mon.a (mon.0) 134 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cluster 2026-03-31T11:33:44.718097+0000 mon.a (mon.0) 134 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.739421+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.739421+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.742544+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.742544+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.747964+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.747964+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.764004+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.764004+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.766569+0000 mon.a (mon.0) 139 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.766569+0000 mon.a (mon.0) 139 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.767150+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.767150+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.772350+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.772350+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.786078+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.786078+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.787519+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.787519+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.787949+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.787949+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.791954+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:44.791954+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.065483+0000 mon.a (mon.0) 146 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.065483+0000 mon.a (mon.0) 146 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.143640+0000 mon.a (mon.0) 147 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.143640+0000 mon.a (mon.0) 147 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.147136+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.147136+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.403759+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.403759+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.407726+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.407726+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.411162+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.565 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.411162+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.511722+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.511722+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.512829+0000 mon.a (mon.0) 153 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.512829+0000 mon.a (mon.0) 153 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.514250+0000 mon.a (mon.0) 154 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.514250+0000 mon.a (mon.0) 154 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.514702+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.514702+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.518256+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.518256+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.519717+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.519717+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.520333+0000 mon.a (mon.0) 158 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.520333+0000 mon.a (mon.0) 158 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.505345+0000 mgr.a (mgr.14152) 27 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm06:192.168.123.106=a;vm09:192.168.123.109=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.505345+0000 mgr.a (mgr.14152) 27 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm06:192.168.123.106=a;vm09:192.168.123.109=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:45.507270+0000 mgr.a (mgr.14152) 28 : cephadm [INF] Saving service mon spec with placement vm06:192.168.123.106=a;vm09:192.168.123.109=b;count:2 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:45.507270+0000 mgr.a (mgr.14152) 28 : cephadm [INF] Saving service mon spec with placement vm06:192.168.123.106=a;vm09:192.168.123.109=b;count:2 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:45.521191+0000 mgr.a (mgr.14152) 29 : cephadm [INF] Deploying daemon mon.b on vm09 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: cephadm 2026-03-31T11:33:45.521191+0000 mgr.a (mgr.14152) 29 : cephadm [INF] Deploying daemon mon.b on vm09 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.859631+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.859631+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.865985+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:45.865985+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:46.370021+0000 mon.a (mon.0) 161 : audit [DBG] from='client.? 192.168.123.109:0/2703158535' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: audit 2026-03-31T11:33:46.370021+0000 mon.a (mon.0) 161 : audit [DBG] from='client.? 192.168.123.109:0/2703158535' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-31T11:33:47.566 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:47 vm09 bash[20661]: debug 2026-03-31T11:33:47.310+0000 7f68dfba1640 1 mon.b@-1(synchronizing).paxosservice(auth 1..5) refresh upgraded, format 0 -> 3 2026-03-31T11:33:47.591 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-31T11:33:47.592 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph mon dump -f json 2026-03-31T11:33:47.900 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.b/config 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cephadm 2026-03-31T11:33:47.201645+0000 mgr.a (mgr.14152) 30 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cephadm 2026-03-31T11:33:47.201645+0000 mgr.a (mgr.14152) 30 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cephadm 2026-03-31T11:33:47.213362+0000 mgr.a (mgr.14152) 31 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cephadm 2026-03-31T11:33:47.213362+0000 mgr.a (mgr.14152) 31 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cephadm 2026-03-31T11:33:47.215252+0000 mgr.a (mgr.14152) 32 : cephadm [INF] Reconfiguring daemon mon.a on vm06 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cephadm 2026-03-31T11:33:47.215252+0000 mgr.a (mgr.14152) 32 : cephadm [INF] Reconfiguring daemon mon.a on vm06 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:47.321719+0000 mon.a (mon.0) 174 : cluster [INF] mon.a calling monitor election 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:47.321719+0000 mon.a (mon.0) 174 : cluster [INF] mon.a calling monitor election 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:47.324639+0000 mon.a (mon.0) 175 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:47.324639+0000 mon.a (mon.0) 175 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:48.319370+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:48.319370+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:49.319546+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:49.319546+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:49.320005+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:49.320005+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:50.319452+0000 mon.a (mon.0) 178 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:50.319452+0000 mon.a (mon.0) 178 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:51.319147+0000 mon.a (mon.0) 179 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:51.319147+0000 mon.a (mon.0) 179 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.319801+0000 mon.a (mon.0) 180 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.319801+0000 mon.a (mon.0) 180 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.327725+0000 mon.a (mon.0) 181 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.327725+0000 mon.a (mon.0) 181 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332718+0000 mon.a (mon.0) 182 : cluster [DBG] monmap epoch 2 2026-03-31T11:33:52.656 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332718+0000 mon.a (mon.0) 182 : cluster [DBG] monmap epoch 2 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332736+0000 mon.a (mon.0) 183 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332736+0000 mon.a (mon.0) 183 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332746+0000 mon.a (mon.0) 184 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332746+0000 mon.a (mon.0) 184 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332757+0000 mon.a (mon.0) 185 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332757+0000 mon.a (mon.0) 185 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332767+0000 mon.a (mon.0) 186 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332767+0000 mon.a (mon.0) 186 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332778+0000 mon.a (mon.0) 187 : cluster [DBG] election_strategy: 1 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332778+0000 mon.a (mon.0) 187 : cluster [DBG] election_strategy: 1 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332788+0000 mon.a (mon.0) 188 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332788+0000 mon.a (mon.0) 188 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332797+0000 mon.a (mon.0) 189 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.332797+0000 mon.a (mon.0) 189 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.333161+0000 mon.a (mon.0) 190 : cluster [DBG] fsmap 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.333161+0000 mon.a (mon.0) 190 : cluster [DBG] fsmap 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.333185+0000 mon.a (mon.0) 191 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.333185+0000 mon.a (mon.0) 191 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.333325+0000 mon.a (mon.0) 192 : cluster [DBG] mgrmap e12: a(active, since 17s) 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.333325+0000 mon.a (mon.0) 192 : cluster [DBG] mgrmap e12: a(active, since 17s) 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.333413+0000 mon.a (mon.0) 193 : cluster [INF] overall HEALTH_OK 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: cluster 2026-03-31T11:33:52.333413+0000 mon.a (mon.0) 193 : cluster [INF] overall HEALTH_OK 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.336712+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.336712+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.340483+0000 mon.a (mon.0) 195 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.340483+0000 mon.a (mon.0) 195 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.343246+0000 mon.a (mon.0) 196 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.343246+0000 mon.a (mon.0) 196 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.345700+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.345700+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.346836+0000 mon.a (mon.0) 198 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.346836+0000 mon.a (mon.0) 198 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.347232+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.347232+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.347663+0000 mon.a (mon.0) 200 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:52.657 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:52 vm09 bash[20661]: audit 2026-03-31T11:33:52.347663+0000 mon.a (mon.0) 200 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:52.712 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:33:52.713 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":2,"fsid":"48134446-2cf5-11f1-830d-4121aa8d72d5","modified":"2026-03-31T11:33:47.318700Z","created":"2026-03-31T11:33:10.045398Z","min_mon_release":20,"min_mon_release_name":"tentacle","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid","tentacle"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-31T11:33:52.713 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 2 2026-03-31T11:33:52.795 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-31T11:33:52.795 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph config generate-minimal-conf 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cephadm 2026-03-31T11:33:47.201645+0000 mgr.a (mgr.14152) 30 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cephadm 2026-03-31T11:33:47.201645+0000 mgr.a (mgr.14152) 30 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cephadm 2026-03-31T11:33:47.213362+0000 mgr.a (mgr.14152) 31 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cephadm 2026-03-31T11:33:47.213362+0000 mgr.a (mgr.14152) 31 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cephadm 2026-03-31T11:33:47.215252+0000 mgr.a (mgr.14152) 32 : cephadm [INF] Reconfiguring daemon mon.a on vm06 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cephadm 2026-03-31T11:33:47.215252+0000 mgr.a (mgr.14152) 32 : cephadm [INF] Reconfiguring daemon mon.a on vm06 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:47.321719+0000 mon.a (mon.0) 174 : cluster [INF] mon.a calling monitor election 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:47.321719+0000 mon.a (mon.0) 174 : cluster [INF] mon.a calling monitor election 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:47.324639+0000 mon.a (mon.0) 175 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:47.324639+0000 mon.a (mon.0) 175 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:48.319370+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:48.319370+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:49.319546+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:49.319546+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:49.320005+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:49.320005+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:50.319452+0000 mon.a (mon.0) 178 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.802 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:50.319452+0000 mon.a (mon.0) 178 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:51.319147+0000 mon.a (mon.0) 179 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:51.319147+0000 mon.a (mon.0) 179 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.319801+0000 mon.a (mon.0) 180 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.319801+0000 mon.a (mon.0) 180 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.327725+0000 mon.a (mon.0) 181 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.327725+0000 mon.a (mon.0) 181 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332718+0000 mon.a (mon.0) 182 : cluster [DBG] monmap epoch 2 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332718+0000 mon.a (mon.0) 182 : cluster [DBG] monmap epoch 2 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332736+0000 mon.a (mon.0) 183 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332736+0000 mon.a (mon.0) 183 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332746+0000 mon.a (mon.0) 184 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332746+0000 mon.a (mon.0) 184 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332757+0000 mon.a (mon.0) 185 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332757+0000 mon.a (mon.0) 185 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332767+0000 mon.a (mon.0) 186 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332767+0000 mon.a (mon.0) 186 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332778+0000 mon.a (mon.0) 187 : cluster [DBG] election_strategy: 1 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332778+0000 mon.a (mon.0) 187 : cluster [DBG] election_strategy: 1 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332788+0000 mon.a (mon.0) 188 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332788+0000 mon.a (mon.0) 188 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332797+0000 mon.a (mon.0) 189 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.332797+0000 mon.a (mon.0) 189 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.333161+0000 mon.a (mon.0) 190 : cluster [DBG] fsmap 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.333161+0000 mon.a (mon.0) 190 : cluster [DBG] fsmap 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.333185+0000 mon.a (mon.0) 191 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.333185+0000 mon.a (mon.0) 191 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.333325+0000 mon.a (mon.0) 192 : cluster [DBG] mgrmap e12: a(active, since 17s) 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.333325+0000 mon.a (mon.0) 192 : cluster [DBG] mgrmap e12: a(active, since 17s) 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.333413+0000 mon.a (mon.0) 193 : cluster [INF] overall HEALTH_OK 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: cluster 2026-03-31T11:33:52.333413+0000 mon.a (mon.0) 193 : cluster [INF] overall HEALTH_OK 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.336712+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.336712+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.340483+0000 mon.a (mon.0) 195 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.340483+0000 mon.a (mon.0) 195 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.343246+0000 mon.a (mon.0) 196 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.343246+0000 mon.a (mon.0) 196 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.345700+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.345700+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.346836+0000 mon.a (mon.0) 198 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.346836+0000 mon.a (mon.0) 198 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.347232+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.347232+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.347663+0000 mon.a (mon.0) 200 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:52.803 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:52 vm06 bash[20457]: audit 2026-03-31T11:33:52.347663+0000 mon.a (mon.0) 200 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.059 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:53.417 INFO:teuthology.orchestra.run.vm06.stdout:# minimal ceph.conf for 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:53.417 INFO:teuthology.orchestra.run.vm06.stdout:[global] 2026-03-31T11:33:53.417 INFO:teuthology.orchestra.run.vm06.stdout: fsid = 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:33:53.417 INFO:teuthology.orchestra.run.vm06.stdout: mon_host = [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-31T11:33:53.498 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-31T11:33:53.498 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:33:53.498 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.conf 2026-03-31T11:33:53.549 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:33:53.549 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:53.601 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:33:53.601 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.conf 2026-03-31T11:33:53.609 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:33:53.609 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.346525+0000 mgr.a (mgr.14152) 33 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.346525+0000 mgr.a (mgr.14152) 33 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.350398+0000 mgr.a (mgr.14152) 34 : cephadm [INF] Reconfiguring daemon mon.b on vm09 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.350398+0000 mgr.a (mgr.14152) 34 : cephadm [INF] Reconfiguring daemon mon.b on vm09 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.712439+0000 mon.a (mon.0) 201 : audit [DBG] from='client.? 192.168.123.109:0/4162330871' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.712439+0000 mon.a (mon.0) 201 : audit [DBG] from='client.? 192.168.123.109:0/4162330871' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.797384+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.797384+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.801918+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.801918+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.803829+0000 mon.a (mon.0) 204 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:53.654 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.803829+0000 mon.a (mon.0) 204 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.814459+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.814459+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.815023+0000 mon.a (mon.0) 206 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.815023+0000 mon.a (mon.0) 206 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.817063+0000 mgr.a (mgr.14152) 35 : cephadm [INF] Updating vm06:/etc/ceph/ceph.conf 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.817063+0000 mgr.a (mgr.14152) 35 : cephadm [INF] Updating vm06:/etc/ceph/ceph.conf 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.817497+0000 mgr.a (mgr.14152) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.817497+0000 mgr.a (mgr.14152) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.906073+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.906073+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.910775+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.910775+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.915732+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.915732+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.920835+0000 mon.a (mon.0) 210 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.920835+0000 mon.a (mon.0) 210 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.924126+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.924126+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.935856+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.935856+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.936456+0000 mon.a (mon.0) 213 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.936456+0000 mon.a (mon.0) 213 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.936877+0000 mon.a (mon.0) 214 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:52.936877+0000 mon.a (mon.0) 214 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:53.319872+0000 mon.a (mon.0) 215 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:53.319872+0000 mon.a (mon.0) 215 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:53.356372+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:53.356372+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:53.360711+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.655 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:53 vm09 bash[20661]: audit 2026-03-31T11:33:53.360711+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.662 INFO:tasks.cephadm:Adding mgr.a on vm06 2026-03-31T11:33:53.663 INFO:tasks.cephadm:Adding mgr.b on vm09 2026-03-31T11:33:53.663 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch apply mgr '2;vm06=a;vm09=b' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.346525+0000 mgr.a (mgr.14152) 33 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.346525+0000 mgr.a (mgr.14152) 33 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.350398+0000 mgr.a (mgr.14152) 34 : cephadm [INF] Reconfiguring daemon mon.b on vm09 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.350398+0000 mgr.a (mgr.14152) 34 : cephadm [INF] Reconfiguring daemon mon.b on vm09 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.712439+0000 mon.a (mon.0) 201 : audit [DBG] from='client.? 192.168.123.109:0/4162330871' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.712439+0000 mon.a (mon.0) 201 : audit [DBG] from='client.? 192.168.123.109:0/4162330871' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.797384+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.797384+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.801918+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.801918+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.803829+0000 mon.a (mon.0) 204 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.803829+0000 mon.a (mon.0) 204 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.814459+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.814459+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.815023+0000 mon.a (mon.0) 206 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.815023+0000 mon.a (mon.0) 206 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.817063+0000 mgr.a (mgr.14152) 35 : cephadm [INF] Updating vm06:/etc/ceph/ceph.conf 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.817063+0000 mgr.a (mgr.14152) 35 : cephadm [INF] Updating vm06:/etc/ceph/ceph.conf 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.817497+0000 mgr.a (mgr.14152) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.817497+0000 mgr.a (mgr.14152) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.906073+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.906073+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.910775+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.910775+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.915732+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.915732+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.920835+0000 mon.a (mon.0) 210 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.920835+0000 mon.a (mon.0) 210 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.924126+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.924126+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.935856+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.935856+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.936456+0000 mon.a (mon.0) 213 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.936456+0000 mon.a (mon.0) 213 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.936877+0000 mon.a (mon.0) 214 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:52.936877+0000 mon.a (mon.0) 214 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:53.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:53.319872+0000 mon.a (mon.0) 215 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:53.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:53.319872+0000 mon.a (mon.0) 215 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:33:53.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:53.356372+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:53.356372+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:53.360711+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:53 vm06 bash[20457]: audit 2026-03-31T11:33:53.360711+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:53.974 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.b/config 2026-03-31T11:33:54.389 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mgr update... 2026-03-31T11:33:54.393 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20727]: debug 2026-03-31T11:33:54.316+0000 7f6f62a60640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-31T11:33:54.465 DEBUG:teuthology.orchestra.run.vm09:mgr.b> sudo journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.b.service 2026-03-31T11:33:54.467 DEBUG:tasks.cephadm:set 0 configs 2026-03-31T11:33:54.467 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph config dump 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.860797+0000 mgr.a (mgr.14152) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.860797+0000 mgr.a (mgr.14152) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.863589+0000 mgr.a (mgr.14152) 38 : cephadm [INF] Updating vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.863589+0000 mgr.a (mgr.14152) 38 : cephadm [INF] Updating vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.924622+0000 mgr.a (mgr.14152) 39 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.924622+0000 mgr.a (mgr.14152) 39 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.935511+0000 mgr.a (mgr.14152) 40 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.935511+0000 mgr.a (mgr.14152) 40 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.937348+0000 mgr.a (mgr.14152) 41 : cephadm [INF] Reconfiguring daemon mon.a on vm06 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: cephadm 2026-03-31T11:33:52.937348+0000 mgr.a (mgr.14152) 41 : cephadm [INF] Reconfiguring daemon mon.a on vm06 2026-03-31T11:33:54.706 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:53.417307+0000 mon.a (mon.0) 218 : audit [DBG] from='client.? 192.168.123.106:0/517085276' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:53.417307+0000 mon.a (mon.0) 218 : audit [DBG] from='client.? 192.168.123.106:0/517085276' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.103804+0000 mon.a (mon.0) 219 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.103804+0000 mon.a (mon.0) 219 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.121173+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.121173+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.127663+0000 mon.a (mon.0) 221 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.127663+0000 mon.a (mon.0) 221 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.129303+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.129303+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.129754+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.129754+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.132668+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.132668+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.137475+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.137475+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.388396+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.388396+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.389261+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 bash[20661]: audit 2026-03-31T11:33:54.389261+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:54.707 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:54 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:54.759 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.860797+0000 mgr.a (mgr.14152) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.860797+0000 mgr.a (mgr.14152) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.863589+0000 mgr.a (mgr.14152) 38 : cephadm [INF] Updating vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.863589+0000 mgr.a (mgr.14152) 38 : cephadm [INF] Updating vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/config/ceph.conf 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.924622+0000 mgr.a (mgr.14152) 39 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.924622+0000 mgr.a (mgr.14152) 39 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.935511+0000 mgr.a (mgr.14152) 40 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.935511+0000 mgr.a (mgr.14152) 40 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.937348+0000 mgr.a (mgr.14152) 41 : cephadm [INF] Reconfiguring daemon mon.a on vm06 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: cephadm 2026-03-31T11:33:52.937348+0000 mgr.a (mgr.14152) 41 : cephadm [INF] Reconfiguring daemon mon.a on vm06 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:53.417307+0000 mon.a (mon.0) 218 : audit [DBG] from='client.? 192.168.123.106:0/517085276' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:53.417307+0000 mon.a (mon.0) 218 : audit [DBG] from='client.? 192.168.123.106:0/517085276' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.103804+0000 mon.a (mon.0) 219 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.103804+0000 mon.a (mon.0) 219 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.121173+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.121173+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.127663+0000 mon.a (mon.0) 221 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.127663+0000 mon.a (mon.0) 221 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.129303+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.129303+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.129754+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.129754+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.132668+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.132668+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.137475+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.137475+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.388396+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.388396+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.389261+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:54.776 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:54 vm06 bash[20457]: audit 2026-03-31T11:33:54.389261+0000 mon.a (mon.0) 227 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:WHO MASK LEVEL OPTION VALUE RO 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global dev auth_debug true 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global basic container_image quay.ceph.io/ceph-ci/ceph@sha256:02c8d616f8a7af1a26efca44a51de7761356c5cb66c69c789e66f798c27c8072 * 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global dev debug_asserts_on_shutdown true 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global basic log_to_file true 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global basic log_to_journald false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global basic log_to_stderr false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_allow_pool_delete true 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_clock_drift_allowed 1.000000 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_cluster_log_to_file true 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_election_default_strategy 3 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_max_pg_per_osd 10000 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_pg_warn_max_object_skew 0.000000 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_warn_on_crush_straw_calc_version_zero false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_warn_on_legacy_crush_tunables false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_warn_on_osd_down_out_interval_zero false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global dev mon_warn_on_pool_pg_num_not_power_of_two false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced mon_warn_on_too_few_osds false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global dev ms_die_on_bug true 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global dev ms_die_on_old_message true 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced osd_pool_default_erasure_code_profile plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced osd_pool_default_pg_autoscale_mode off 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:global advanced public_network 192.168.123.0/24,192.168.123.1/32 * 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced auth_allow_insecure_global_id_reclaim false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced auth_mon_ticket_ttl 660.000000 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced auth_service_ticket_ttl 240.000000 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced debug_mon 20/20 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced debug_ms 1/1 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced debug_paxos 20/20 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced mon_data_avail_warn 5 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced mon_mgr_mkfs_grace 240 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon dev mon_osd_prime_pg_temp true 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced mon_osd_reporter_subtree_level osd 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced mon_reweight_min_bytes_per_osd 10 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced mon_reweight_min_pgs_per_osd 4 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced mon_warn_on_insecure_global_id_reclaim false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mon advanced mon_warn_on_insecure_global_id_reclaim_allowed false 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced debug_mgr 20/20 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced debug_ms 1/1 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced mgr/cephadm/allow_ptrace true * 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced mgr/cephadm/container_init True * 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced mgr/cephadm/migration_current 7 * 2026-03-31T11:33:55.094 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced mgr/cephadm/use_agent true * 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced mgr/dashboard/ssl_server_port 8443 * 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced mgr/orchestrator/orchestrator cephadm 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced mon_reweight_min_bytes_per_osd 10 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:mgr advanced mon_reweight_min_pgs_per_osd 4 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev bdev_debug_aio true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced debug_ms 1/1 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced debug_osd 20/20 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev osd_debug_misdirected_ops true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev osd_debug_op_order true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev osd_debug_pg_log_writeout true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev osd_debug_shutdown true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev osd_debug_verify_cached_snaps true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev osd_debug_verify_missing_on_start true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev osd_debug_verify_stray_on_activate true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_deep_scrub_update_digest_min_age 30 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd basic osd_mclock_iops_capacity_threshold_hdd 49000.000000 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_mclock_profile high_recovery_ops 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd dev osd_mclock_skip_benchmark true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_memory_target_autotune true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_op_queue debug_random * 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_op_queue_cut_off debug_random * 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_recover_clone_overlap true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_recovery_max_chunk 1048576 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_scrub_load_threshold 5.000000 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_scrub_max_interval 600.000000 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:osd advanced osd_shutdown_pgref_assert true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:client.rgw advanced rgw_cache_enabled true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:client.rgw advanced rgw_enable_ops_log true 2026-03-31T11:33:55.095 INFO:teuthology.orchestra.run.vm06.stdout:client.rgw advanced rgw_enable_usage_log true 2026-03-31T11:33:55.163 INFO:tasks.cephadm:Deploying OSDs... 2026-03-31T11:33:55.163 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-31T11:33:55.163 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-03-31T11:33:55.166 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-31T11:33:55.166 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_1 2026-03-31T11:33:55.211 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-31T11:33:55.212 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-31T11:33:55.212 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 776 Links: 1 2026-03-31T11:33:55.212 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-31T11:33:55.212 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-31 11:30:56.477125000 +0000 2026-03-31T11:33:55.212 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-31 11:30:56.353125000 +0000 2026-03-31T11:33:55.212 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-31 11:30:56.353125000 +0000 2026-03-31T11:33:55.212 INFO:teuthology.orchestra.run.vm06.stdout: Birth: - 2026-03-31T11:33:55.212 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-31T11:33:55.261 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-31T11:33:55.261 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-31T11:33:55.261 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000182682 s, 2.8 MB/s 2026-03-31T11:33:55.261 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-31T11:33:55.292 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:54 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:55.292 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:55.292 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:55 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:55.292 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:55 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:55.292 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:55 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:55.292 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:55 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:33:55.309 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_2 2026-03-31T11:33:55.359 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-31T11:33:55.359 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-31T11:33:55.359 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 808 Links: 1 2026-03-31T11:33:55.359 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-31T11:33:55.359 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-31 11:30:56.625125000 +0000 2026-03-31T11:33:55.359 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-31 11:30:56.621125000 +0000 2026-03-31T11:33:55.359 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-31 11:30:56.621125000 +0000 2026-03-31T11:33:55.359 INFO:teuthology.orchestra.run.vm06.stdout: Birth: - 2026-03-31T11:33:55.359 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-31T11:33:55.411 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-31T11:33:55.411 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-31T11:33:55.411 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000147566 s, 3.5 MB/s 2026-03-31T11:33:55.412 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: cluster 2026-03-31T11:33:54.357879+0000 mgr.a (mgr.14152) 42 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: cluster 2026-03-31T11:33:54.357879+0000 mgr.a (mgr.14152) 42 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.382144+0000 mgr.a (mgr.14152) 43 : audit [DBG] from='client.14200 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm06=a;vm09=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.382144+0000 mgr.a (mgr.14152) 43 : audit [DBG] from='client.14200 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm06=a;vm09=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: cephadm 2026-03-31T11:33:54.383145+0000 mgr.a (mgr.14152) 44 : cephadm [INF] Saving service mgr spec with placement vm06=a;vm09=b;count:2 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: cephadm 2026-03-31T11:33:54.383145+0000 mgr.a (mgr.14152) 44 : cephadm [INF] Saving service mgr spec with placement vm06=a;vm09=b;count:2 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.390994+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.390994+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.391447+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.391447+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.403946+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.403946+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.405280+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.405280+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.408018+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.408018+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.410243+0000 mon.a (mon.0) 233 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr services"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.410243+0000 mon.a (mon.0) 233 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr services"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.410868+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:54.410868+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: cephadm 2026-03-31T11:33:54.411497+0000 mgr.a (mgr.14152) 45 : cephadm [INF] Deploying daemon mgr.b on vm09 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: cephadm 2026-03-31T11:33:54.411497+0000 mgr.a (mgr.14152) 45 : cephadm [INF] Deploying daemon mgr.b on vm09 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.093826+0000 mon.a (mon.0) 235 : audit [DBG] from='client.? 192.168.123.106:0/974187889' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-31T11:33:55.412 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.093826+0000 mon.a (mon.0) 235 : audit [DBG] from='client.? 192.168.123.106:0/974187889' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.323982+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.323982+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.326870+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.326870+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.329593+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.329593+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.332880+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.332880+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.344100+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.344100+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.345296+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.345296+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.345696+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.345696+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.349560+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.349560+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.359960+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.359960+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.360585+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr services"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.360585+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr services"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.361003+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.413 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:55 vm06 bash[20457]: audit 2026-03-31T11:33:55.361003+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.417 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_3 2026-03-31T11:33:55.464 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-31T11:33:55.464 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-31T11:33:55.464 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 840 Links: 1 2026-03-31T11:33:55.464 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-31T11:33:55.464 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-31 11:30:57.049125000 +0000 2026-03-31T11:33:55.464 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-31 11:30:56.921125000 +0000 2026-03-31T11:33:55.464 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-31 11:30:56.921125000 +0000 2026-03-31T11:33:55.464 INFO:teuthology.orchestra.run.vm06.stdout: Birth: - 2026-03-31T11:33:55.464 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-31T11:33:55.517 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-31T11:33:55.517 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-31T11:33:55.517 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000202759 s, 2.5 MB/s 2026-03-31T11:33:55.518 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-31T11:33:55.566 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_4 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: cluster 2026-03-31T11:33:54.357879+0000 mgr.a (mgr.14152) 42 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: cluster 2026-03-31T11:33:54.357879+0000 mgr.a (mgr.14152) 42 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.382144+0000 mgr.a (mgr.14152) 43 : audit [DBG] from='client.14200 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm06=a;vm09=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.382144+0000 mgr.a (mgr.14152) 43 : audit [DBG] from='client.14200 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm06=a;vm09=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: cephadm 2026-03-31T11:33:54.383145+0000 mgr.a (mgr.14152) 44 : cephadm [INF] Saving service mgr spec with placement vm06=a;vm09=b;count:2 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: cephadm 2026-03-31T11:33:54.383145+0000 mgr.a (mgr.14152) 44 : cephadm [INF] Saving service mgr spec with placement vm06=a;vm09=b;count:2 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.390994+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.390994+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.391447+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.391447+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.403946+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.403946+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.405280+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.405280+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.408018+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.408018+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.410243+0000 mon.a (mon.0) 233 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr services"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.410243+0000 mon.a (mon.0) 233 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr services"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.410868+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:54.410868+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: cephadm 2026-03-31T11:33:54.411497+0000 mgr.a (mgr.14152) 45 : cephadm [INF] Deploying daemon mgr.b on vm09 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: cephadm 2026-03-31T11:33:54.411497+0000 mgr.a (mgr.14152) 45 : cephadm [INF] Deploying daemon mgr.b on vm09 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.093826+0000 mon.a (mon.0) 235 : audit [DBG] from='client.? 192.168.123.106:0/974187889' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.093826+0000 mon.a (mon.0) 235 : audit [DBG] from='client.? 192.168.123.106:0/974187889' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.323982+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.323982+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.326870+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.326870+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.329593+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.329593+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.332880+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.332880+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.344100+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.344100+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.345296+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.345296+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.345696+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.345696+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.349560+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.349560+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.359960+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.359960+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.360585+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr services"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.360585+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr services"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.361003+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[20661]: audit 2026-03-31T11:33:55.361003+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:55 vm09 systemd[1]: Started Ceph mgr.b for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:33:55.586 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:55 vm09 bash[21551]: debug 2026-03-31T11:33:55.402+0000 7f9386e8e640 1 -- 192.168.123.109:0/2818760978 <== mon.0 v2:192.168.123.106:3300/0 4 ==== auth_reply(proto 2 0 (0) Success) ==== 194+0+0 (secure 0 0 0) 0x5612b0d38000 con 0x5612b0d3b400 2026-03-31T11:33:55.617 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-31T11:33:55.617 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-31T11:33:55.617 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 873 Links: 1 2026-03-31T11:33:55.617 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-31T11:33:55.617 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-31 11:31:00.881125000 +0000 2026-03-31T11:33:55.617 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-31 11:30:57.213125000 +0000 2026-03-31T11:33:55.617 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-31 11:30:57.213125000 +0000 2026-03-31T11:33:55.617 INFO:teuthology.orchestra.run.vm06.stdout: Birth: - 2026-03-31T11:33:55.617 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-31T11:33:55.674 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-31T11:33:55.674 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-31T11:33:55.674 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000134451 s, 3.8 MB/s 2026-03-31T11:33:55.674 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-31T11:33:55.726 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-31T11:33:55.726 DEBUG:teuthology.orchestra.run.vm09:> dd if=/scratch_devs of=/dev/stdout 2026-03-31T11:33:55.730 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-31T11:33:55.731 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vg_nvme/lv_1 2026-03-31T11:33:55.777 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-31T11:33:55.777 INFO:teuthology.orchestra.run.vm09.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-31T11:33:55.777 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 784 Links: 1 2026-03-31T11:33:55.777 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-31T11:33:55.777 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-31 11:30:54.331550000 +0000 2026-03-31T11:33:55.777 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-31 11:30:54.327550000 +0000 2026-03-31T11:33:55.777 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-31 11:30:54.327550000 +0000 2026-03-31T11:33:55.777 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-31T11:33:55.777 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-31T11:33:55.825 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-31T11:33:55.825 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-31T11:33:55.825 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000161422 s, 3.2 MB/s 2026-03-31T11:33:55.825 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-31T11:33:55.875 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vg_nvme/lv_2 2026-03-31T11:33:55.921 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-31T11:33:55.921 INFO:teuthology.orchestra.run.vm09.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-31T11:33:55.921 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 816 Links: 1 2026-03-31T11:33:55.921 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-31T11:33:55.921 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-31 11:30:54.619550000 +0000 2026-03-31T11:33:55.921 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-31 11:30:54.615550000 +0000 2026-03-31T11:33:55.921 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-31 11:30:54.615550000 +0000 2026-03-31T11:33:55.921 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-31T11:33:55.921 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-31T11:33:55.969 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-31T11:33:55.969 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-31T11:33:55.969 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000111918 s, 4.6 MB/s 2026-03-31T11:33:55.970 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-31T11:33:56.023 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vg_nvme/lv_3 2026-03-31T11:33:56.073 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-31T11:33:56.073 INFO:teuthology.orchestra.run.vm09.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-31T11:33:56.073 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 846 Links: 1 2026-03-31T11:33:56.073 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-31T11:33:56.073 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-31 11:30:55.031550000 +0000 2026-03-31T11:33:56.073 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-31 11:30:54.899550000 +0000 2026-03-31T11:33:56.073 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-31 11:30:54.899550000 +0000 2026-03-31T11:33:56.073 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-31T11:33:56.073 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-31T11:33:56.121 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-31T11:33:56.121 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-31T11:33:56.122 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000179626 s, 2.9 MB/s 2026-03-31T11:33:56.122 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-31T11:33:56.171 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vg_nvme/lv_4 2026-03-31T11:33:56.217 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-31T11:33:56.217 INFO:teuthology.orchestra.run.vm09.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-31T11:33:56.217 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 880 Links: 1 2026-03-31T11:33:56.217 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-31T11:33:56.217 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-31 11:31:00.871550000 +0000 2026-03-31T11:33:56.217 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-31 11:30:55.191550000 +0000 2026-03-31T11:33:56.217 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-31 11:30:55.191550000 +0000 2026-03-31T11:33:56.217 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-31T11:33:56.217 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-31T11:33:56.266 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-31T11:33:56.266 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-31T11:33:56.267 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.00019778 s, 2.6 MB/s 2026-03-31T11:33:56.267 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-31T11:33:56.315 INFO:tasks.cephadm:Deploying osd.0 on vm06 with /dev/vg_nvme/lv_4... 2026-03-31T11:33:56.315 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- lvm zap /dev/vg_nvme/lv_4 2026-03-31T11:33:56.614 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: cephadm 2026-03-31T11:33:55.349885+0000 mgr.a (mgr.14152) 46 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: cephadm 2026-03-31T11:33:55.349885+0000 mgr.a (mgr.14152) 46 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: cephadm 2026-03-31T11:33:55.359669+0000 mgr.a (mgr.14152) 47 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: cephadm 2026-03-31T11:33:55.359669+0000 mgr.a (mgr.14152) 47 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: cephadm 2026-03-31T11:33:55.361447+0000 mgr.a (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mgr.a on vm06 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: cephadm 2026-03-31T11:33:55.361447+0000 mgr.a (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mgr.a on vm06 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.817469+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.817469+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.821980+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.821980+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.823368+0000 mon.a (mon.0) 249 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.823368+0000 mon.a (mon.0) 249 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.824636+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.824636+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.825120+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.825120+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.830077+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:55.830077+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:56.724153+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:56.885 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[20661]: audit 2026-03-31T11:33:56.724153+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: cephadm 2026-03-31T11:33:55.349885+0000 mgr.a (mgr.14152) 46 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: cephadm 2026-03-31T11:33:55.349885+0000 mgr.a (mgr.14152) 46 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: cephadm 2026-03-31T11:33:55.359669+0000 mgr.a (mgr.14152) 47 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: cephadm 2026-03-31T11:33:55.359669+0000 mgr.a (mgr.14152) 47 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: cephadm 2026-03-31T11:33:55.361447+0000 mgr.a (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mgr.a on vm06 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: cephadm 2026-03-31T11:33:55.361447+0000 mgr.a (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mgr.a on vm06 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.817469+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.817469+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.821980+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.821980+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.823368+0000 mon.a (mon.0) 249 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.823368+0000 mon.a (mon.0) 249 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.824636+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.824636+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.825120+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.825120+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.830077+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:55.830077+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:56.724153+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:56 vm06 bash[20457]: audit 2026-03-31T11:33:56.724153+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:57.199 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:33:57.232 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch daemon add osd vm06:vg_nvme/lv_4 --skip-validation 2026-03-31T11:33:57.334 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[21551]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-31T11:33:57.334 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[21551]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-31T11:33:57.334 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:33:56 vm09 bash[21551]: from numpy import show_config as show_numpy_config 2026-03-31T11:33:57.494 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:33:58.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:57 vm09 bash[20661]: cephadm 2026-03-31T11:33:55.830564+0000 mgr.a (mgr.14152) 49 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:58.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:57 vm09 bash[20661]: cephadm 2026-03-31T11:33:55.830564+0000 mgr.a (mgr.14152) 49 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:58.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:57 vm09 bash[20661]: cluster 2026-03-31T11:33:56.358116+0000 mgr.a (mgr.14152) 50 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:58.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:57 vm09 bash[20661]: cluster 2026-03-31T11:33:56.358116+0000 mgr.a (mgr.14152) 50 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:58.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:57 vm09 bash[20661]: audit 2026-03-31T11:33:57.347585+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:58.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:57 vm09 bash[20661]: audit 2026-03-31T11:33:57.347585+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:58.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:57 vm09 bash[20661]: audit 2026-03-31T11:33:57.818001+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:58.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:57 vm09 bash[20661]: audit 2026-03-31T11:33:57.818001+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:58.139 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:57 vm06 bash[20457]: cephadm 2026-03-31T11:33:55.830564+0000 mgr.a (mgr.14152) 49 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:58.139 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:57 vm06 bash[20457]: cephadm 2026-03-31T11:33:55.830564+0000 mgr.a (mgr.14152) 49 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:58.139 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:57 vm06 bash[20457]: cluster 2026-03-31T11:33:56.358116+0000 mgr.a (mgr.14152) 50 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:58.139 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:57 vm06 bash[20457]: cluster 2026-03-31T11:33:56.358116+0000 mgr.a (mgr.14152) 50 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:58.139 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:57 vm06 bash[20457]: audit 2026-03-31T11:33:57.347585+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:58.139 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:57 vm06 bash[20457]: audit 2026-03-31T11:33:57.347585+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:58.139 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:57 vm06 bash[20457]: audit 2026-03-31T11:33:57.818001+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:58.139 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:57 vm06 bash[20457]: audit 2026-03-31T11:33:57.818001+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.812367+0000 mgr.a (mgr.14152) 51 : audit [DBG] from='client.14212 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.812367+0000 mgr.a (mgr.14152) 51 : audit [DBG] from='client.14212 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.813356+0000 mgr.a (mgr.14152) 52 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.813356+0000 mgr.a (mgr.14152) 52 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.813452+0000 mgr.a (mgr.14152) 53 : cephadm [INF] Creating OSDs with service ID: default on vm06:['vg_nvme/lv_4'] 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.813452+0000 mgr.a (mgr.14152) 53 : cephadm [INF] Creating OSDs with service ID: default on vm06:['vg_nvme/lv_4'] 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.818671+0000 mgr.a (mgr.14152) 54 : cephadm [INF] Marking host: vm06 for OSDSpec preview refresh. 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.818671+0000 mgr.a (mgr.14152) 54 : cephadm [INF] Marking host: vm06 for OSDSpec preview refresh. 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.818875+0000 mgr.a (mgr.14152) 55 : cephadm [INF] Saving service osd.default spec with placement vm06 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.818875+0000 mgr.a (mgr.14152) 55 : cephadm [INF] Saving service osd.default spec with placement vm06 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.822885+0000 mon.a (mon.0) 256 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.822885+0000 mon.a (mon.0) 256 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.823949+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.823949+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:33:59.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.824074+0000 mon.a (mon.0) 258 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.824074+0000 mon.a (mon.0) 258 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.828853+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.828853+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.829385+0000 mon.a (mon.0) 260 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.829385+0000 mon.a (mon.0) 260 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.832584+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.832584+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.839700+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.839700+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.841545+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.841545+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.842015+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.842015+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.849640+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:57.849640+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.850208+0000 mgr.a (mgr.14152) 56 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cephadm 2026-03-31T11:33:57.850208+0000 mgr.a (mgr.14152) 56 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cluster 2026-03-31T11:33:58.358311+0000 mgr.a (mgr.14152) 57 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cluster 2026-03-31T11:33:58.358311+0000 mgr.a (mgr.14152) 57 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:58.775607+0000 mon.a (mon.0) 266 : audit [INF] from='client.? 192.168.123.106:0/66844800' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "88f60178-5fdf-4f09-a5f8-805f65d0338d"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:58.775607+0000 mon.a (mon.0) 266 : audit [INF] from='client.? 192.168.123.106:0/66844800' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "88f60178-5fdf-4f09-a5f8-805f65d0338d"} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:58.779669+0000 mon.a (mon.0) 267 : audit [INF] from='client.? 192.168.123.106:0/66844800' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "88f60178-5fdf-4f09-a5f8-805f65d0338d"}]': finished 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:58.779669+0000 mon.a (mon.0) 267 : audit [INF] from='client.? 192.168.123.106:0/66844800' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "88f60178-5fdf-4f09-a5f8-805f65d0338d"}]': finished 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cluster 2026-03-31T11:33:58.783117+0000 mon.a (mon.0) 268 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: cluster 2026-03-31T11:33:58.783117+0000 mon.a (mon.0) 268 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:58.783413+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:33:59.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:58 vm09 bash[20661]: audit 2026-03-31T11:33:58.783413+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.812367+0000 mgr.a (mgr.14152) 51 : audit [DBG] from='client.14212 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.812367+0000 mgr.a (mgr.14152) 51 : audit [DBG] from='client.14212 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.813356+0000 mgr.a (mgr.14152) 52 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.813356+0000 mgr.a (mgr.14152) 52 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.813452+0000 mgr.a (mgr.14152) 53 : cephadm [INF] Creating OSDs with service ID: default on vm06:['vg_nvme/lv_4'] 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.813452+0000 mgr.a (mgr.14152) 53 : cephadm [INF] Creating OSDs with service ID: default on vm06:['vg_nvme/lv_4'] 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.818671+0000 mgr.a (mgr.14152) 54 : cephadm [INF] Marking host: vm06 for OSDSpec preview refresh. 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.818671+0000 mgr.a (mgr.14152) 54 : cephadm [INF] Marking host: vm06 for OSDSpec preview refresh. 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.818875+0000 mgr.a (mgr.14152) 55 : cephadm [INF] Saving service osd.default spec with placement vm06 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.818875+0000 mgr.a (mgr.14152) 55 : cephadm [INF] Saving service osd.default spec with placement vm06 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.822885+0000 mon.a (mon.0) 256 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.822885+0000 mon.a (mon.0) 256 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.823949+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.823949+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.824074+0000 mon.a (mon.0) 258 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.824074+0000 mon.a (mon.0) 258 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.828853+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.828853+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.829385+0000 mon.a (mon.0) 260 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.829385+0000 mon.a (mon.0) 260 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.832584+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.832584+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.839700+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.839700+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.841545+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.841545+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.842015+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.842015+0000 mon.a (mon.0) 264 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.849640+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:57.849640+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.850208+0000 mgr.a (mgr.14152) 56 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cephadm 2026-03-31T11:33:57.850208+0000 mgr.a (mgr.14152) 56 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cluster 2026-03-31T11:33:58.358311+0000 mgr.a (mgr.14152) 57 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cluster 2026-03-31T11:33:58.358311+0000 mgr.a (mgr.14152) 57 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:58.775607+0000 mon.a (mon.0) 266 : audit [INF] from='client.? 192.168.123.106:0/66844800' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "88f60178-5fdf-4f09-a5f8-805f65d0338d"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:58.775607+0000 mon.a (mon.0) 266 : audit [INF] from='client.? 192.168.123.106:0/66844800' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "88f60178-5fdf-4f09-a5f8-805f65d0338d"} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:58.779669+0000 mon.a (mon.0) 267 : audit [INF] from='client.? 192.168.123.106:0/66844800' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "88f60178-5fdf-4f09-a5f8-805f65d0338d"}]': finished 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:58.779669+0000 mon.a (mon.0) 267 : audit [INF] from='client.? 192.168.123.106:0/66844800' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "88f60178-5fdf-4f09-a5f8-805f65d0338d"}]': finished 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cluster 2026-03-31T11:33:58.783117+0000 mon.a (mon.0) 268 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: cluster 2026-03-31T11:33:58.783117+0000 mon.a (mon.0) 268 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:58.783413+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:33:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:58 vm06 bash[20457]: audit 2026-03-31T11:33:58.783413+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: cluster 2026-03-31T11:33:59.020912+0000 mon.a (mon.0) 270 : cluster [DBG] Standby manager daemon b started 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: cluster 2026-03-31T11:33:59.020912+0000 mon.a (mon.0) 270 : cluster [DBG] Standby manager daemon b started 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.022851+0000 mon.a (mon.0) 271 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/b/crt"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.022851+0000 mon.a (mon.0) 271 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/b/crt"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.023210+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/crt"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.023210+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/crt"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.023803+0000 mon.a (mon.0) 273 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/b/key"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.023803+0000 mon.a (mon.0) 273 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/b/key"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.024049+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/key"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.024049+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/key"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.137974+0000 mon.a (mon.0) 275 : audit [DBG] from='client.? 192.168.123.106:0/2919364549' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-31T11:34:00.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:33:59 vm06 bash[20457]: audit 2026-03-31T11:33:59.137974+0000 mon.a (mon.0) 275 : audit [DBG] from='client.? 192.168.123.106:0/2919364549' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: cluster 2026-03-31T11:33:59.020912+0000 mon.a (mon.0) 270 : cluster [DBG] Standby manager daemon b started 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: cluster 2026-03-31T11:33:59.020912+0000 mon.a (mon.0) 270 : cluster [DBG] Standby manager daemon b started 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.022851+0000 mon.a (mon.0) 271 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/b/crt"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.022851+0000 mon.a (mon.0) 271 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/b/crt"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.023210+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/crt"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.023210+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/crt"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.023803+0000 mon.a (mon.0) 273 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/b/key"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.023803+0000 mon.a (mon.0) 273 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/b/key"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.024049+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/key"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.024049+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.? 192.168.123.109:0/1370947411' entity='mgr.b' cmd={"prefix": "config-key get", "key": "mgr/dashboard/key"} : dispatch 2026-03-31T11:34:00.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.137974+0000 mon.a (mon.0) 275 : audit [DBG] from='client.? 192.168.123.106:0/2919364549' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-31T11:34:00.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:33:59 vm09 bash[20661]: audit 2026-03-31T11:33:59.137974+0000 mon.a (mon.0) 275 : audit [DBG] from='client.? 192.168.123.106:0/2919364549' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-31T11:34:01.216 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:00 vm06 bash[20457]: cluster 2026-03-31T11:33:59.866215+0000 mon.a (mon.0) 276 : cluster [DBG] mgrmap e13: a(active, since 25s), standbys: b 2026-03-31T11:34:01.216 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:00 vm06 bash[20457]: cluster 2026-03-31T11:33:59.866215+0000 mon.a (mon.0) 276 : cluster [DBG] mgrmap e13: a(active, since 25s), standbys: b 2026-03-31T11:34:01.216 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:00 vm06 bash[20457]: audit 2026-03-31T11:33:59.866343+0000 mon.a (mon.0) 277 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "b", "id": "b"} : dispatch 2026-03-31T11:34:01.216 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:00 vm06 bash[20457]: audit 2026-03-31T11:33:59.866343+0000 mon.a (mon.0) 277 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "b", "id": "b"} : dispatch 2026-03-31T11:34:01.216 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:00 vm06 bash[20457]: cluster 2026-03-31T11:34:00.358494+0000 mgr.a (mgr.14152) 58 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:01.216 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:00 vm06 bash[20457]: cluster 2026-03-31T11:34:00.358494+0000 mgr.a (mgr.14152) 58 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:01.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:00 vm09 bash[20661]: cluster 2026-03-31T11:33:59.866215+0000 mon.a (mon.0) 276 : cluster [DBG] mgrmap e13: a(active, since 25s), standbys: b 2026-03-31T11:34:01.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:00 vm09 bash[20661]: cluster 2026-03-31T11:33:59.866215+0000 mon.a (mon.0) 276 : cluster [DBG] mgrmap e13: a(active, since 25s), standbys: b 2026-03-31T11:34:01.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:00 vm09 bash[20661]: audit 2026-03-31T11:33:59.866343+0000 mon.a (mon.0) 277 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "b", "id": "b"} : dispatch 2026-03-31T11:34:01.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:00 vm09 bash[20661]: audit 2026-03-31T11:33:59.866343+0000 mon.a (mon.0) 277 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mgr metadata", "who": "b", "id": "b"} : dispatch 2026-03-31T11:34:01.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:00 vm09 bash[20661]: cluster 2026-03-31T11:34:00.358494+0000 mgr.a (mgr.14152) 58 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:01.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:00 vm09 bash[20661]: cluster 2026-03-31T11:34:00.358494+0000 mgr.a (mgr.14152) 58 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:01.871 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:34:01 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:01.871 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:01 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:01.871 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:01 vm06 bash[20457]: audit 2026-03-31T11:34:00.913407+0000 mon.a (mon.0) 278 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-31T11:34:01.871 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:01 vm06 bash[20457]: audit 2026-03-31T11:34:00.913407+0000 mon.a (mon.0) 278 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-31T11:34:01.871 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:01 vm06 bash[20457]: audit 2026-03-31T11:34:00.914012+0000 mon.a (mon.0) 279 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:01.871 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:01 vm06 bash[20457]: audit 2026-03-31T11:34:00.914012+0000 mon.a (mon.0) 279 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:01.871 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:01 vm06 bash[20457]: cephadm 2026-03-31T11:34:00.914472+0000 mgr.a (mgr.14152) 59 : cephadm [INF] Deploying daemon osd.0 on vm06 2026-03-31T11:34:01.871 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:01 vm06 bash[20457]: cephadm 2026-03-31T11:34:00.914472+0000 mgr.a (mgr.14152) 59 : cephadm [INF] Deploying daemon osd.0 on vm06 2026-03-31T11:34:02.129 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:02.129 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:34:02 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:02.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:01 vm09 bash[20661]: audit 2026-03-31T11:34:00.913407+0000 mon.a (mon.0) 278 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-31T11:34:02.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:01 vm09 bash[20661]: audit 2026-03-31T11:34:00.913407+0000 mon.a (mon.0) 278 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-31T11:34:02.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:01 vm09 bash[20661]: audit 2026-03-31T11:34:00.914012+0000 mon.a (mon.0) 279 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:02.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:01 vm09 bash[20661]: audit 2026-03-31T11:34:00.914012+0000 mon.a (mon.0) 279 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:02.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:01 vm09 bash[20661]: cephadm 2026-03-31T11:34:00.914472+0000 mgr.a (mgr.14152) 59 : cephadm [INF] Deploying daemon osd.0 on vm06 2026-03-31T11:34:02.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:01 vm09 bash[20661]: cephadm 2026-03-31T11:34:00.914472+0000 mgr.a (mgr.14152) 59 : cephadm [INF] Deploying daemon osd.0 on vm06 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.126351+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.126351+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.133150+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.133150+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.136099+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.136099+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.137841+0000 mon.a (mon.0) 283 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.137841+0000 mon.a (mon.0) 283 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.142448+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.142448+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.146614+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.146614+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: cephadm 2026-03-31T11:34:02.149049+0000 mgr.a (mgr.14152) 60 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: cephadm 2026-03-31T11:34:02.149049+0000 mgr.a (mgr.14152) 60 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: cluster 2026-03-31T11:34:02.358703+0000 mgr.a (mgr.14152) 61 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: cluster 2026-03-31T11:34:02.358703+0000 mgr.a (mgr.14152) 61 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.585063+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.585063+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.020 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.591199+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.591199+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.594346+0000 mon.a (mon.0) 288 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.594346+0000 mon.a (mon.0) 288 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.598375+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.598375+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.598643+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.598643+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.600460+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.600460+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.607181+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.607181+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.609043+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.609043+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.612154+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.612154+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.612740+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.021 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:02 vm06 bash[20457]: audit 2026-03-31T11:34:02.612740+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.126351+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.126351+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.133150+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.133150+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.136099+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.136099+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.137841+0000 mon.a (mon.0) 283 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.137841+0000 mon.a (mon.0) 283 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.142448+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.142448+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.146614+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.146614+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: cephadm 2026-03-31T11:34:02.149049+0000 mgr.a (mgr.14152) 60 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: cephadm 2026-03-31T11:34:02.149049+0000 mgr.a (mgr.14152) 60 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: cluster 2026-03-31T11:34:02.358703+0000 mgr.a (mgr.14152) 61 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: cluster 2026-03-31T11:34:02.358703+0000 mgr.a (mgr.14152) 61 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.585063+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.585063+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.591199+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.591199+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.594346+0000 mon.a (mon.0) 288 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.594346+0000 mon.a (mon.0) 288 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.598375+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.598375+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.598643+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.598643+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.600460+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.600460+0000 mon.a (mon.0) 291 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.607181+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.607181+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.609043+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.609043+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.612154+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.612154+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.612740+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:03.336 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:02 vm09 bash[20661]: audit 2026-03-31T11:34:02.612740+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:04.243 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 0 on host 'vm06' 2026-03-31T11:34:04.338 DEBUG:teuthology.orchestra.run.vm06:osd.0> sudo journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@osd.0.service 2026-03-31T11:34:04.339 INFO:tasks.cephadm:Deploying osd.1 on vm09 with /dev/vg_nvme/lv_4... 2026-03-31T11:34:04.339 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- lvm zap /dev/vg_nvme/lv_4 2026-03-31T11:34:04.658 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.b/config 2026-03-31T11:34:04.816 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:34:04 vm06 bash[28816]: debug 2026-03-31T11:34:04.372+0000 7f9d02b348c0 -1 osd.0 0 log_to_monitors true 2026-03-31T11:34:05.262 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-31T11:34:05.292 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch daemon add osd vm09:vg_nvme/lv_4 --skip-validation 2026-03-31T11:34:05.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:05 vm06 bash[20457]: audit 2026-03-31T11:34:04.236896+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:05.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:05 vm06 bash[20457]: audit 2026-03-31T11:34:04.236896+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:05.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:05 vm06 bash[20457]: audit 2026-03-31T11:34:04.242018+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:05.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:05 vm06 bash[20457]: audit 2026-03-31T11:34:04.242018+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:05.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:05 vm06 bash[20457]: cluster 2026-03-31T11:34:04.359012+0000 mgr.a (mgr.14152) 62 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:05.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:05 vm06 bash[20457]: cluster 2026-03-31T11:34:04.359012+0000 mgr.a (mgr.14152) 62 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:05.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:05 vm06 bash[20457]: audit 2026-03-31T11:34:04.375393+0000 mon.a (mon.0) 298 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-31T11:34:05.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:05 vm06 bash[20457]: audit 2026-03-31T11:34:04.375393+0000 mon.a (mon.0) 298 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-31T11:34:05.567 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.b/config 2026-03-31T11:34:05.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:05 vm09 bash[20661]: audit 2026-03-31T11:34:04.236896+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:05.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:05 vm09 bash[20661]: audit 2026-03-31T11:34:04.236896+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:05.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:05 vm09 bash[20661]: audit 2026-03-31T11:34:04.242018+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:05.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:05 vm09 bash[20661]: audit 2026-03-31T11:34:04.242018+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:05.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:05 vm09 bash[20661]: cluster 2026-03-31T11:34:04.359012+0000 mgr.a (mgr.14152) 62 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:05.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:05 vm09 bash[20661]: cluster 2026-03-31T11:34:04.359012+0000 mgr.a (mgr.14152) 62 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:05.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:05 vm09 bash[20661]: audit 2026-03-31T11:34:04.375393+0000 mon.a (mon.0) 298 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-31T11:34:05.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:05 vm09 bash[20661]: audit 2026-03-31T11:34:04.375393+0000 mon.a (mon.0) 298 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-31T11:34:06.565 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:34:06 vm06 bash[28816]: debug 2026-03-31T11:34:06.264+0000 7f9cfeab6640 -1 osd.0 0 waiting for initial osdmap 2026-03-31T11:34:06.565 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:34:06 vm06 bash[28816]: debug 2026-03-31T11:34:06.264+0000 7f9cf9885640 -1 osd.0 7 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.244114+0000 mon.a (mon.0) 299 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.244114+0000 mon.a (mon.0) 299 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: cluster 2026-03-31T11:34:05.246704+0000 mon.a (mon.0) 300 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: cluster 2026-03-31T11:34:05.246704+0000 mon.a (mon.0) 300 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.247348+0000 mon.a (mon.0) 301 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.247348+0000 mon.a (mon.0) 301 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.247466+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.247466+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.481581+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.481581+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.485837+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.485837+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.486488+0000 mgr.a (mgr.14152) 63 : cephadm [INF] Detected new or changed devices on vm06 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.486488+0000 mgr.a (mgr.14152) 63 : cephadm [INF] Detected new or changed devices on vm06 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.491445+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.491445+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.917170+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.917170+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.921081+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.921081+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.921903+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.921903+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.923568+0000 mon.a (mon.0) 309 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.923568+0000 mon.a (mon.0) 309 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.924251+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:06.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:06 vm06 bash[20457]: audit 2026-03-31T11:34:05.924251+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.244114+0000 mon.a (mon.0) 299 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.244114+0000 mon.a (mon.0) 299 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: cluster 2026-03-31T11:34:05.246704+0000 mon.a (mon.0) 300 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: cluster 2026-03-31T11:34:05.246704+0000 mon.a (mon.0) 300 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.247348+0000 mon.a (mon.0) 301 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.247348+0000 mon.a (mon.0) 301 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.247466+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.247466+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.481581+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.481581+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.485837+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.485837+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.486488+0000 mgr.a (mgr.14152) 63 : cephadm [INF] Detected new or changed devices on vm06 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.486488+0000 mgr.a (mgr.14152) 63 : cephadm [INF] Detected new or changed devices on vm06 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.491445+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.491445+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.917170+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.917170+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.921081+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.921081+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.921903+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.921903+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.923568+0000 mon.a (mon.0) 309 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.923568+0000 mon.a (mon.0) 309 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.924251+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:06.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:06 vm09 bash[20661]: audit 2026-03-31T11:34:05.924251+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:05.910078+0000 mgr.a (mgr.14152) 64 : audit [DBG] from='client.24113 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:05.910078+0000 mgr.a (mgr.14152) 64 : audit [DBG] from='client.24113 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.912353+0000 mgr.a (mgr.14152) 65 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.912353+0000 mgr.a (mgr.14152) 65 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.912423+0000 mgr.a (mgr.14152) 66 : cephadm [INF] Creating OSDs with service ID: default on vm09:['vg_nvme/lv_4'] 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.912423+0000 mgr.a (mgr.14152) 66 : cephadm [INF] Creating OSDs with service ID: default on vm09:['vg_nvme/lv_4'] 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.917808+0000 mgr.a (mgr.14152) 67 : cephadm [INF] Marking host: vm09 for OSDSpec preview refresh. 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.917808+0000 mgr.a (mgr.14152) 67 : cephadm [INF] Marking host: vm09 for OSDSpec preview refresh. 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.917913+0000 mgr.a (mgr.14152) 68 : cephadm [INF] Saving service osd.default spec with placement vm09 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cephadm 2026-03-31T11:34:05.917913+0000 mgr.a (mgr.14152) 68 : cephadm [INF] Saving service osd.default spec with placement vm09 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.246299+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.246299+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cluster 2026-03-31T11:34:06.248337+0000 mon.a (mon.0) 312 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cluster 2026-03-31T11:34:06.248337+0000 mon.a (mon.0) 312 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.249326+0000 mon.a (mon.0) 313 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.249326+0000 mon.a (mon.0) 313 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.253801+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.253801+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cluster 2026-03-31T11:34:06.359189+0000 mgr.a (mgr.14152) 69 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cluster 2026-03-31T11:34:06.359189+0000 mgr.a (mgr.14152) 69 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.775933+0000 mon.b (mon.1) 2 : audit [INF] from='client.? 192.168.123.109:0/2745881167' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.775933+0000 mon.b (mon.1) 2 : audit [INF] from='client.? 192.168.123.109:0/2745881167' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.776531+0000 mon.a (mon.0) 315 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.776531+0000 mon.a (mon.0) 315 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"} : dispatch 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.779345+0000 mon.a (mon.0) 316 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"}]': finished 2026-03-31T11:34:07.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.779345+0000 mon.a (mon.0) 316 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"}]': finished 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cluster 2026-03-31T11:34:06.781468+0000 mon.a (mon.0) 317 : cluster [INF] osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751] boot 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cluster 2026-03-31T11:34:06.781468+0000 mon.a (mon.0) 317 : cluster [INF] osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751] boot 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cluster 2026-03-31T11:34:06.781481+0000 mon.a (mon.0) 318 : cluster [DBG] osdmap e8: 2 total, 1 up, 2 in 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: cluster 2026-03-31T11:34:06.781481+0000 mon.a (mon.0) 318 : cluster [DBG] osdmap e8: 2 total, 1 up, 2 in 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.781623+0000 mon.a (mon.0) 319 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.781623+0000 mon.a (mon.0) 319 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.781757+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.781757+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.996038+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.996038+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.998750+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:06.998750+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.009440+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.009440+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.011511+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.011511+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.014136+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.014136+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.017114+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.017114+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.017566+0000 mon.a (mon.0) 327 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.017566+0000 mon.a (mon.0) 327 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.017957+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.017957+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.020543+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.020543+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.021402+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.021402+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.022410+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.022410+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.023099+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.023099+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.124473+0000 mon.b (mon.1) 3 : audit [DBG] from='client.? 192.168.123.109:0/1564539727' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-31T11:34:07.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:07 vm06 bash[20457]: audit 2026-03-31T11:34:07.124473+0000 mon.b (mon.1) 3 : audit [DBG] from='client.? 192.168.123.109:0/1564539727' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:05.910078+0000 mgr.a (mgr.14152) 64 : audit [DBG] from='client.24113 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:05.910078+0000 mgr.a (mgr.14152) 64 : audit [DBG] from='client.24113 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.912353+0000 mgr.a (mgr.14152) 65 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.912353+0000 mgr.a (mgr.14152) 65 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.912423+0000 mgr.a (mgr.14152) 66 : cephadm [INF] Creating OSDs with service ID: default on vm09:['vg_nvme/lv_4'] 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.912423+0000 mgr.a (mgr.14152) 66 : cephadm [INF] Creating OSDs with service ID: default on vm09:['vg_nvme/lv_4'] 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.917808+0000 mgr.a (mgr.14152) 67 : cephadm [INF] Marking host: vm09 for OSDSpec preview refresh. 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.917808+0000 mgr.a (mgr.14152) 67 : cephadm [INF] Marking host: vm09 for OSDSpec preview refresh. 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.917913+0000 mgr.a (mgr.14152) 68 : cephadm [INF] Saving service osd.default spec with placement vm09 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cephadm 2026-03-31T11:34:05.917913+0000 mgr.a (mgr.14152) 68 : cephadm [INF] Saving service osd.default spec with placement vm09 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.246299+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.246299+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cluster 2026-03-31T11:34:06.248337+0000 mon.a (mon.0) 312 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cluster 2026-03-31T11:34:06.248337+0000 mon.a (mon.0) 312 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.249326+0000 mon.a (mon.0) 313 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.249326+0000 mon.a (mon.0) 313 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.253801+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.253801+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cluster 2026-03-31T11:34:06.359189+0000 mgr.a (mgr.14152) 69 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cluster 2026-03-31T11:34:06.359189+0000 mgr.a (mgr.14152) 69 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.775933+0000 mon.b (mon.1) 2 : audit [INF] from='client.? 192.168.123.109:0/2745881167' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.775933+0000 mon.b (mon.1) 2 : audit [INF] from='client.? 192.168.123.109:0/2745881167' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.776531+0000 mon.a (mon.0) 315 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.776531+0000 mon.a (mon.0) 315 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"} : dispatch 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.779345+0000 mon.a (mon.0) 316 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"}]': finished 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.779345+0000 mon.a (mon.0) 316 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "dabbc60e-788e-440b-9f82-29039e83287a"}]': finished 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cluster 2026-03-31T11:34:06.781468+0000 mon.a (mon.0) 317 : cluster [INF] osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751] boot 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cluster 2026-03-31T11:34:06.781468+0000 mon.a (mon.0) 317 : cluster [INF] osd.0 [v2:192.168.123.106:6802/3317286751,v1:192.168.123.106:6803/3317286751] boot 2026-03-31T11:34:07.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cluster 2026-03-31T11:34:06.781481+0000 mon.a (mon.0) 318 : cluster [DBG] osdmap e8: 2 total, 1 up, 2 in 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: cluster 2026-03-31T11:34:06.781481+0000 mon.a (mon.0) 318 : cluster [DBG] osdmap e8: 2 total, 1 up, 2 in 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.781623+0000 mon.a (mon.0) 319 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.781623+0000 mon.a (mon.0) 319 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.781757+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.781757+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.996038+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.996038+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.998750+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:06.998750+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.009440+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.009440+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.011511+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.011511+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.014136+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.014136+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.017114+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.017114+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.017566+0000 mon.a (mon.0) 327 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.017566+0000 mon.a (mon.0) 327 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.017957+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.017957+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.020543+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.020543+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.021402+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.021402+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.022410+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.022410+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.023099+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.023099+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.124473+0000 mon.b (mon.1) 3 : audit [DBG] from='client.? 192.168.123.109:0/1564539727' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-31T11:34:07.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:07 vm09 bash[20661]: audit 2026-03-31T11:34:07.124473+0000 mon.b (mon.1) 3 : audit [DBG] from='client.? 192.168.123.109:0/1564539727' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-31T11:34:08.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:08 vm06 bash[20457]: cluster 2026-03-31T11:34:05.374878+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-31T11:34:08.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:08 vm06 bash[20457]: cluster 2026-03-31T11:34:05.374878+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-31T11:34:08.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:08 vm06 bash[20457]: cluster 2026-03-31T11:34:05.374889+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-31T11:34:08.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:08 vm06 bash[20457]: cluster 2026-03-31T11:34:05.374889+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-31T11:34:08.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:08 vm09 bash[20661]: cluster 2026-03-31T11:34:05.374878+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-31T11:34:08.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:08 vm09 bash[20661]: cluster 2026-03-31T11:34:05.374878+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-31T11:34:08.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:08 vm09 bash[20661]: cluster 2026-03-31T11:34:05.374889+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-31T11:34:08.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:08 vm09 bash[20661]: cluster 2026-03-31T11:34:05.374889+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: cluster 2026-03-31T11:34:08.266896+0000 mon.a (mon.0) 333 : cluster [DBG] osdmap e9: 2 total, 1 up, 2 in 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: cluster 2026-03-31T11:34:08.266896+0000 mon.a (mon.0) 333 : cluster [DBG] osdmap e9: 2 total, 1 up, 2 in 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:08.267280+0000 mon.a (mon.0) 334 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:08.267280+0000 mon.a (mon.0) 334 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: cluster 2026-03-31T11:34:08.359386+0000 mgr.a (mgr.14152) 70 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: cluster 2026-03-31T11:34:08.359386+0000 mgr.a (mgr.14152) 70 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:08.684087+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:08.684087+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:08.684558+0000 mon.a (mon.0) 336 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:08.684558+0000 mon.a (mon.0) 336 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: cephadm 2026-03-31T11:34:08.684916+0000 mgr.a (mgr.14152) 71 : cephadm [INF] Deploying daemon osd.1 on vm09 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: cephadm 2026-03-31T11:34:08.684916+0000 mgr.a (mgr.14152) 71 : cephadm [INF] Deploying daemon osd.1 on vm09 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:09.027996+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:09.027996+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:09.028580+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:09.531 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 bash[20661]: audit 2026-03-31T11:34:09.028580+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:09.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: cluster 2026-03-31T11:34:08.266896+0000 mon.a (mon.0) 333 : cluster [DBG] osdmap e9: 2 total, 1 up, 2 in 2026-03-31T11:34:09.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: cluster 2026-03-31T11:34:08.266896+0000 mon.a (mon.0) 333 : cluster [DBG] osdmap e9: 2 total, 1 up, 2 in 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:08.267280+0000 mon.a (mon.0) 334 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:08.267280+0000 mon.a (mon.0) 334 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: cluster 2026-03-31T11:34:08.359386+0000 mgr.a (mgr.14152) 70 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: cluster 2026-03-31T11:34:08.359386+0000 mgr.a (mgr.14152) 70 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:08.684087+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:08.684087+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:08.684558+0000 mon.a (mon.0) 336 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:08.684558+0000 mon.a (mon.0) 336 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: cephadm 2026-03-31T11:34:08.684916+0000 mgr.a (mgr.14152) 71 : cephadm [INF] Deploying daemon osd.1 on vm09 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: cephadm 2026-03-31T11:34:08.684916+0000 mgr.a (mgr.14152) 71 : cephadm [INF] Deploying daemon osd.1 on vm09 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:09.027996+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:09.027996+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:09.028580+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:09.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:09 vm06 bash[20457]: audit 2026-03-31T11:34:09.028580+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:34:09.808 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:09.809 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:09 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:09.809 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:34:09 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:09.809 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:34:09 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: cephadm 2026-03-31T11:34:09.029023+0000 mgr.a (mgr.14152) 72 : cephadm [INF] Deploying daemon osd.1 on vm09 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: cephadm 2026-03-31T11:34:09.029023+0000 mgr.a (mgr.14152) 72 : cephadm [INF] Deploying daemon osd.1 on vm09 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:09.833894+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:09.833894+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:09.837732+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:09.837732+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:10.057644+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:10.057644+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:10.103178+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:10.103178+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:10.132500+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.530 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 bash[20661]: audit 2026-03-31T11:34:10.132500+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: cephadm 2026-03-31T11:34:09.029023+0000 mgr.a (mgr.14152) 72 : cephadm [INF] Deploying daemon osd.1 on vm09 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: cephadm 2026-03-31T11:34:09.029023+0000 mgr.a (mgr.14152) 72 : cephadm [INF] Deploying daemon osd.1 on vm09 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:09.833894+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:09.833894+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:09.837732+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:09.837732+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:10.057644+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:10.057644+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:10.103178+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:10.103178+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:10.132500+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:10 vm06 bash[20457]: audit 2026-03-31T11:34:10.132500+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:10.793 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:10 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:10.793 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:34:10 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:34:11.432 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:11 vm09 bash[20661]: cluster 2026-03-31T11:34:10.359732+0000 mgr.a (mgr.14152) 73 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:11.432 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:11 vm09 bash[20661]: cluster 2026-03-31T11:34:10.359732+0000 mgr.a (mgr.14152) 73 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:11.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:11 vm06 bash[20457]: cluster 2026-03-31T11:34:10.359732+0000 mgr.a (mgr.14152) 73 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:11.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:11 vm06 bash[20457]: cluster 2026-03-31T11:34:10.359732+0000 mgr.a (mgr.14152) 73 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:12.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:12 vm06 bash[20457]: audit 2026-03-31T11:34:11.968538+0000 mon.b (mon.1) 4 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:34:12.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:12 vm06 bash[20457]: audit 2026-03-31T11:34:11.968538+0000 mon.b (mon.1) 4 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:34:12.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:12 vm06 bash[20457]: audit 2026-03-31T11:34:11.969080+0000 mon.a (mon.0) 344 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:34:12.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:12 vm06 bash[20457]: audit 2026-03-31T11:34:11.969080+0000 mon.a (mon.0) 344 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:34:12.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:12 vm09 bash[20661]: audit 2026-03-31T11:34:11.968538+0000 mon.b (mon.1) 4 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:34:12.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:12 vm09 bash[20661]: audit 2026-03-31T11:34:11.968538+0000 mon.b (mon.1) 4 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:34:12.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:12 vm09 bash[20661]: audit 2026-03-31T11:34:11.969080+0000 mon.a (mon.0) 344 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:34:12.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:12 vm09 bash[20661]: audit 2026-03-31T11:34:11.969080+0000 mon.a (mon.0) 344 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:12.284593+0000 mon.a (mon.0) 345 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:12.284593+0000 mon.a (mon.0) 345 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: cluster 2026-03-31T11:34:12.286628+0000 mon.a (mon.0) 346 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: cluster 2026-03-31T11:34:12.286628+0000 mon.a (mon.0) 346 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:12.287094+0000 mon.b (mon.1) 5 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:12.287094+0000 mon.b (mon.1) 5 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:12.287549+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:12.287549+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:12.287661+0000 mon.a (mon.0) 348 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:12.287661+0000 mon.a (mon.0) 348 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: cluster 2026-03-31T11:34:12.360080+0000 mgr.a (mgr.14152) 74 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: cluster 2026-03-31T11:34:12.360080+0000 mgr.a (mgr.14152) 74 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:13.287197+0000 mon.a (mon.0) 349 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: audit 2026-03-31T11:34:13.287197+0000 mon.a (mon.0) 349 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: cluster 2026-03-31T11:34:13.289135+0000 mon.a (mon.0) 350 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-31T11:34:13.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:13 vm06 bash[20457]: cluster 2026-03-31T11:34:13.289135+0000 mon.a (mon.0) 350 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-31T11:34:13.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:12.284593+0000 mon.a (mon.0) 345 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-31T11:34:13.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:12.284593+0000 mon.a (mon.0) 345 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-31T11:34:13.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: cluster 2026-03-31T11:34:12.286628+0000 mon.a (mon.0) 346 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-31T11:34:13.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: cluster 2026-03-31T11:34:12.286628+0000 mon.a (mon.0) 346 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-31T11:34:13.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:12.287094+0000 mon.b (mon.1) 5 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:34:13.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:12.287094+0000 mon.b (mon.1) 5 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:12.287549+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:12.287549+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:12.287661+0000 mon.a (mon.0) 348 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:12.287661+0000 mon.a (mon.0) 348 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: cluster 2026-03-31T11:34:12.360080+0000 mgr.a (mgr.14152) 74 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: cluster 2026-03-31T11:34:12.360080+0000 mgr.a (mgr.14152) 74 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:13.287197+0000 mon.a (mon.0) 349 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: audit 2026-03-31T11:34:13.287197+0000 mon.a (mon.0) 349 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: cluster 2026-03-31T11:34:13.289135+0000 mon.a (mon.0) 350 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-31T11:34:13.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:13 vm09 bash[20661]: cluster 2026-03-31T11:34:13.289135+0000 mon.a (mon.0) 350 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-31T11:34:14.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:14 vm06 bash[20457]: audit 2026-03-31T11:34:13.289810+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:14.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:14 vm06 bash[20457]: audit 2026-03-31T11:34:13.289810+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:14.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:14 vm06 bash[20457]: audit 2026-03-31T11:34:13.294948+0000 mon.a (mon.0) 352 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:14.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:14 vm06 bash[20457]: audit 2026-03-31T11:34:13.294948+0000 mon.a (mon.0) 352 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:14.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:14 vm09 bash[20661]: audit 2026-03-31T11:34:13.289810+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:14.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:14 vm09 bash[20661]: audit 2026-03-31T11:34:13.289810+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:14.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:14 vm09 bash[20661]: audit 2026-03-31T11:34:13.294948+0000 mon.a (mon.0) 352 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:14.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:14 vm09 bash[20661]: audit 2026-03-31T11:34:13.294948+0000 mon.a (mon.0) 352 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:12.966586+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:12.966586+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:12.966601+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:12.966601+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: audit 2026-03-31T11:34:14.295861+0000 mon.a (mon.0) 353 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: audit 2026-03-31T11:34:14.295861+0000 mon.a (mon.0) 353 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:14.300168+0000 mon.a (mon.0) 354 : cluster [INF] osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058] boot 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:14.300168+0000 mon.a (mon.0) 354 : cluster [INF] osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058] boot 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:14.300201+0000 mon.a (mon.0) 355 : cluster [DBG] osdmap e12: 2 total, 2 up, 2 in 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:14.300201+0000 mon.a (mon.0) 355 : cluster [DBG] osdmap e12: 2 total, 2 up, 2 in 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: audit 2026-03-31T11:34:14.301049+0000 mon.a (mon.0) 356 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: audit 2026-03-31T11:34:14.301049+0000 mon.a (mon.0) 356 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:14.360483+0000 mgr.a (mgr.14152) 75 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:15.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:15 vm06 bash[20457]: cluster 2026-03-31T11:34:14.360483+0000 mgr.a (mgr.14152) 75 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:12.966586+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:12.966586+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:12.966601+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:12.966601+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: audit 2026-03-31T11:34:14.295861+0000 mon.a (mon.0) 353 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: audit 2026-03-31T11:34:14.295861+0000 mon.a (mon.0) 353 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:14.300168+0000 mon.a (mon.0) 354 : cluster [INF] osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058] boot 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:14.300168+0000 mon.a (mon.0) 354 : cluster [INF] osd.1 [v2:192.168.123.109:6800/3356105058,v1:192.168.123.109:6801/3356105058] boot 2026-03-31T11:34:15.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:14.300201+0000 mon.a (mon.0) 355 : cluster [DBG] osdmap e12: 2 total, 2 up, 2 in 2026-03-31T11:34:15.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:14.300201+0000 mon.a (mon.0) 355 : cluster [DBG] osdmap e12: 2 total, 2 up, 2 in 2026-03-31T11:34:15.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: audit 2026-03-31T11:34:14.301049+0000 mon.a (mon.0) 356 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: audit 2026-03-31T11:34:14.301049+0000 mon.a (mon.0) 356 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:34:15.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:14.360483+0000 mgr.a (mgr.14152) 75 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:15.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:15 vm09 bash[20661]: cluster 2026-03-31T11:34:14.360483+0000 mgr.a (mgr.14152) 75 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-31T11:34:16.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:16 vm06 bash[20457]: cluster 2026-03-31T11:34:15.304317+0000 mon.a (mon.0) 357 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-31T11:34:16.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:16 vm06 bash[20457]: cluster 2026-03-31T11:34:15.304317+0000 mon.a (mon.0) 357 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-31T11:34:16.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:16 vm09 bash[20661]: cluster 2026-03-31T11:34:15.304317+0000 mon.a (mon.0) 357 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-31T11:34:16.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:16 vm09 bash[20661]: cluster 2026-03-31T11:34:15.304317+0000 mon.a (mon.0) 357 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-31T11:34:17.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:17 vm06 bash[20457]: cluster 2026-03-31T11:34:16.360732+0000 mgr.a (mgr.14152) 76 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:17.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:17 vm06 bash[20457]: cluster 2026-03-31T11:34:16.360732+0000 mgr.a (mgr.14152) 76 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:17.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:17 vm09 bash[20661]: cluster 2026-03-31T11:34:16.360732+0000 mgr.a (mgr.14152) 76 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:17.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:17 vm09 bash[20661]: cluster 2026-03-31T11:34:16.360732+0000 mgr.a (mgr.14152) 76 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:19.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:19 vm06 bash[20457]: cluster 2026-03-31T11:34:18.360912+0000 mgr.a (mgr.14152) 77 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:19.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:19 vm06 bash[20457]: cluster 2026-03-31T11:34:18.360912+0000 mgr.a (mgr.14152) 77 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:19.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:19 vm09 bash[20661]: cluster 2026-03-31T11:34:18.360912+0000 mgr.a (mgr.14152) 77 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:19.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:19 vm09 bash[20661]: cluster 2026-03-31T11:34:18.360912+0000 mgr.a (mgr.14152) 77 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:21.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:21 vm06 bash[20457]: cluster 2026-03-31T11:34:20.361101+0000 mgr.a (mgr.14152) 78 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:21.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:21 vm06 bash[20457]: cluster 2026-03-31T11:34:20.361101+0000 mgr.a (mgr.14152) 78 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:21.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:21 vm09 bash[20661]: cluster 2026-03-31T11:34:20.361101+0000 mgr.a (mgr.14152) 78 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:21.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:21 vm09 bash[20661]: cluster 2026-03-31T11:34:20.361101+0000 mgr.a (mgr.14152) 78 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:24.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:23 vm06 bash[20457]: cluster 2026-03-31T11:34:22.361310+0000 mgr.a (mgr.14152) 79 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:24.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:23 vm06 bash[20457]: cluster 2026-03-31T11:34:22.361310+0000 mgr.a (mgr.14152) 79 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:24.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:23 vm06 bash[20457]: audit 2026-03-31T11:34:22.658178+0000 mon.a (mon.0) 358 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:23 vm06 bash[20457]: audit 2026-03-31T11:34:22.658178+0000 mon.a (mon.0) 358 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:23 vm06 bash[20457]: audit 2026-03-31T11:34:22.661369+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:23 vm06 bash[20457]: audit 2026-03-31T11:34:22.661369+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:23 vm06 bash[20457]: audit 2026-03-31T11:34:22.664355+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:23 vm06 bash[20457]: audit 2026-03-31T11:34:22.664355+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:23 vm09 bash[20661]: cluster 2026-03-31T11:34:22.361310+0000 mgr.a (mgr.14152) 79 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:24.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:23 vm09 bash[20661]: cluster 2026-03-31T11:34:22.361310+0000 mgr.a (mgr.14152) 79 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:24.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:23 vm09 bash[20661]: audit 2026-03-31T11:34:22.658178+0000 mon.a (mon.0) 358 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:23 vm09 bash[20661]: audit 2026-03-31T11:34:22.658178+0000 mon.a (mon.0) 358 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:23 vm09 bash[20661]: audit 2026-03-31T11:34:22.661369+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:23 vm09 bash[20661]: audit 2026-03-31T11:34:22.661369+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:23 vm09 bash[20661]: audit 2026-03-31T11:34:22.664355+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:24.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:23 vm09 bash[20661]: audit 2026-03-31T11:34:22.664355+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:25 vm06 bash[20457]: cluster 2026-03-31T11:34:24.361536+0000 mgr.a (mgr.14152) 80 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:26.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:25 vm06 bash[20457]: cluster 2026-03-31T11:34:24.361536+0000 mgr.a (mgr.14152) 80 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:26.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:25 vm06 bash[20457]: audit 2026-03-31T11:34:25.171538+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:25 vm06 bash[20457]: audit 2026-03-31T11:34:25.171538+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:25 vm06 bash[20457]: audit 2026-03-31T11:34:25.174592+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:25 vm06 bash[20457]: audit 2026-03-31T11:34:25.174592+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:25 vm06 bash[20457]: audit 2026-03-31T11:34:25.177932+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:25 vm06 bash[20457]: audit 2026-03-31T11:34:25.177932+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:25 vm09 bash[20661]: cluster 2026-03-31T11:34:24.361536+0000 mgr.a (mgr.14152) 80 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:26.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:25 vm09 bash[20661]: cluster 2026-03-31T11:34:24.361536+0000 mgr.a (mgr.14152) 80 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:26.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:25 vm09 bash[20661]: audit 2026-03-31T11:34:25.171538+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:25 vm09 bash[20661]: audit 2026-03-31T11:34:25.171538+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:25 vm09 bash[20661]: audit 2026-03-31T11:34:25.174592+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:25 vm09 bash[20661]: audit 2026-03-31T11:34:25.174592+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:25 vm09 bash[20661]: audit 2026-03-31T11:34:25.177932+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:25 vm09 bash[20661]: audit 2026-03-31T11:34:25.177932+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: cluster 2026-03-31T11:34:25.175073+0000 mgr.a (mgr.14152) 81 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: cluster 2026-03-31T11:34:25.175073+0000 mgr.a (mgr.14152) 81 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: cluster 2026-03-31T11:34:26.173894+0000 mon.a (mon.0) 364 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: cluster 2026-03-31T11:34:26.173894+0000 mon.a (mon.0) 364 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: audit 2026-03-31T11:34:26.553459+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: audit 2026-03-31T11:34:26.553459+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: audit 2026-03-31T11:34:26.556672+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: audit 2026-03-31T11:34:26.556672+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: audit 2026-03-31T11:34:26.560500+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:26 vm06 bash[20457]: audit 2026-03-31T11:34:26.560500+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:27.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: cluster 2026-03-31T11:34:25.175073+0000 mgr.a (mgr.14152) 81 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:27.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: cluster 2026-03-31T11:34:25.175073+0000 mgr.a (mgr.14152) 81 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:27.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: cluster 2026-03-31T11:34:26.173894+0000 mon.a (mon.0) 364 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-31T11:34:27.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: cluster 2026-03-31T11:34:26.173894+0000 mon.a (mon.0) 364 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-31T11:34:27.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: audit 2026-03-31T11:34:26.553459+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:27.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: audit 2026-03-31T11:34:26.553459+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:27.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: audit 2026-03-31T11:34:26.556672+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:27.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: audit 2026-03-31T11:34:26.556672+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:27.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: audit 2026-03-31T11:34:26.560500+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:27.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:26 vm09 bash[20661]: audit 2026-03-31T11:34:26.560500+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:28.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:27 vm06 bash[20457]: cluster 2026-03-31T11:34:26.557320+0000 mgr.a (mgr.14152) 82 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:28.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:27 vm06 bash[20457]: cluster 2026-03-31T11:34:26.557320+0000 mgr.a (mgr.14152) 82 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:28.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:27 vm06 bash[20457]: cluster 2026-03-31T11:34:26.557416+0000 mgr.a (mgr.14152) 83 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:28.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:27 vm06 bash[20457]: cluster 2026-03-31T11:34:26.557416+0000 mgr.a (mgr.14152) 83 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:28.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:27 vm09 bash[20661]: cluster 2026-03-31T11:34:26.557320+0000 mgr.a (mgr.14152) 82 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:28.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:27 vm09 bash[20661]: cluster 2026-03-31T11:34:26.557320+0000 mgr.a (mgr.14152) 82 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:28.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:27 vm09 bash[20661]: cluster 2026-03-31T11:34:26.557416+0000 mgr.a (mgr.14152) 83 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:28.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:27 vm09 bash[20661]: cluster 2026-03-31T11:34:26.557416+0000 mgr.a (mgr.14152) 83 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:30.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:29 vm06 bash[20457]: cluster 2026-03-31T11:34:28.557623+0000 mgr.a (mgr.14152) 84 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:30.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:29 vm06 bash[20457]: cluster 2026-03-31T11:34:28.557623+0000 mgr.a (mgr.14152) 84 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:30.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:29 vm09 bash[20661]: cluster 2026-03-31T11:34:28.557623+0000 mgr.a (mgr.14152) 84 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:30.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:29 vm09 bash[20661]: cluster 2026-03-31T11:34:28.557623+0000 mgr.a (mgr.14152) 84 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:32.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:31 vm06 bash[20457]: cluster 2026-03-31T11:34:30.557798+0000 mgr.a (mgr.14152) 85 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:32.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:31 vm06 bash[20457]: cluster 2026-03-31T11:34:30.557798+0000 mgr.a (mgr.14152) 85 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:32.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:31 vm09 bash[20661]: cluster 2026-03-31T11:34:30.557798+0000 mgr.a (mgr.14152) 85 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:32.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:31 vm09 bash[20661]: cluster 2026-03-31T11:34:30.557798+0000 mgr.a (mgr.14152) 85 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:34.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:33 vm06 bash[20457]: cluster 2026-03-31T11:34:32.558020+0000 mgr.a (mgr.14152) 86 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:34.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:33 vm06 bash[20457]: cluster 2026-03-31T11:34:32.558020+0000 mgr.a (mgr.14152) 86 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:34.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:33 vm09 bash[20661]: cluster 2026-03-31T11:34:32.558020+0000 mgr.a (mgr.14152) 86 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:34.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:33 vm09 bash[20661]: cluster 2026-03-31T11:34:32.558020+0000 mgr.a (mgr.14152) 86 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:36.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:35 vm06 bash[20457]: cluster 2026-03-31T11:34:34.558229+0000 mgr.a (mgr.14152) 87 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:36.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:35 vm06 bash[20457]: cluster 2026-03-31T11:34:34.558229+0000 mgr.a (mgr.14152) 87 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:36.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:35 vm09 bash[20661]: cluster 2026-03-31T11:34:34.558229+0000 mgr.a (mgr.14152) 87 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:36.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:35 vm09 bash[20661]: cluster 2026-03-31T11:34:34.558229+0000 mgr.a (mgr.14152) 87 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:38.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:37 vm06 bash[20457]: cluster 2026-03-31T11:34:36.558421+0000 mgr.a (mgr.14152) 88 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:38.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:37 vm06 bash[20457]: cluster 2026-03-31T11:34:36.558421+0000 mgr.a (mgr.14152) 88 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:38.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:37 vm09 bash[20661]: cluster 2026-03-31T11:34:36.558421+0000 mgr.a (mgr.14152) 88 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:38.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:37 vm09 bash[20661]: cluster 2026-03-31T11:34:36.558421+0000 mgr.a (mgr.14152) 88 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:40.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:39 vm06 bash[20457]: cluster 2026-03-31T11:34:38.558549+0000 mgr.a (mgr.14152) 89 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:40.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:39 vm06 bash[20457]: cluster 2026-03-31T11:34:38.558549+0000 mgr.a (mgr.14152) 89 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:40.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:39 vm09 bash[20661]: cluster 2026-03-31T11:34:38.558549+0000 mgr.a (mgr.14152) 89 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:40.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:39 vm09 bash[20661]: cluster 2026-03-31T11:34:38.558549+0000 mgr.a (mgr.14152) 89 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:42.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:41 vm06 bash[20457]: cluster 2026-03-31T11:34:40.558688+0000 mgr.a (mgr.14152) 90 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:42.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:41 vm06 bash[20457]: cluster 2026-03-31T11:34:40.558688+0000 mgr.a (mgr.14152) 90 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:42.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:41 vm09 bash[20661]: cluster 2026-03-31T11:34:40.558688+0000 mgr.a (mgr.14152) 90 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:42.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:41 vm09 bash[20661]: cluster 2026-03-31T11:34:40.558688+0000 mgr.a (mgr.14152) 90 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:43.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:42 vm06 bash[20457]: audit 2026-03-31T11:34:41.721807+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:42 vm06 bash[20457]: audit 2026-03-31T11:34:41.721807+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:42 vm06 bash[20457]: audit 2026-03-31T11:34:41.724818+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:42 vm06 bash[20457]: audit 2026-03-31T11:34:41.724818+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:42 vm06 bash[20457]: cluster 2026-03-31T11:34:41.725556+0000 mgr.a (mgr.14152) 91 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:43.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:42 vm06 bash[20457]: cluster 2026-03-31T11:34:41.725556+0000 mgr.a (mgr.14152) 91 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:43.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:42 vm06 bash[20457]: audit 2026-03-31T11:34:41.728090+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:42 vm06 bash[20457]: audit 2026-03-31T11:34:41.728090+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:42 vm09 bash[20661]: audit 2026-03-31T11:34:41.721807+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:42 vm09 bash[20661]: audit 2026-03-31T11:34:41.721807+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:42 vm09 bash[20661]: audit 2026-03-31T11:34:41.724818+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:42 vm09 bash[20661]: audit 2026-03-31T11:34:41.724818+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:42 vm09 bash[20661]: cluster 2026-03-31T11:34:41.725556+0000 mgr.a (mgr.14152) 91 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:43.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:42 vm09 bash[20661]: cluster 2026-03-31T11:34:41.725556+0000 mgr.a (mgr.14152) 91 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:43.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:42 vm09 bash[20661]: audit 2026-03-31T11:34:41.728090+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:43.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:42 vm09 bash[20661]: audit 2026-03-31T11:34:41.728090+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:45.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:44 vm06 bash[20457]: cluster 2026-03-31T11:34:43.725825+0000 mgr.a (mgr.14152) 92 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:45.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:44 vm06 bash[20457]: cluster 2026-03-31T11:34:43.725825+0000 mgr.a (mgr.14152) 92 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:45.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:44 vm09 bash[20661]: cluster 2026-03-31T11:34:43.725825+0000 mgr.a (mgr.14152) 92 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:45.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:44 vm09 bash[20661]: cluster 2026-03-31T11:34:43.725825+0000 mgr.a (mgr.14152) 92 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:47.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:46 vm06 bash[20457]: cluster 2026-03-31T11:34:45.726076+0000 mgr.a (mgr.14152) 93 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:47.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:46 vm06 bash[20457]: cluster 2026-03-31T11:34:45.726076+0000 mgr.a (mgr.14152) 93 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:47.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:46 vm06 bash[20457]: audit 2026-03-31T11:34:46.619997+0000 mon.a (mon.0) 371 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:46 vm06 bash[20457]: audit 2026-03-31T11:34:46.619997+0000 mon.a (mon.0) 371 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:46 vm06 bash[20457]: audit 2026-03-31T11:34:46.622926+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:46 vm06 bash[20457]: audit 2026-03-31T11:34:46.622926+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:46 vm06 bash[20457]: audit 2026-03-31T11:34:46.626262+0000 mon.a (mon.0) 373 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:46 vm06 bash[20457]: audit 2026-03-31T11:34:46.626262+0000 mon.a (mon.0) 373 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:46 vm09 bash[20661]: cluster 2026-03-31T11:34:45.726076+0000 mgr.a (mgr.14152) 93 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:47.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:46 vm09 bash[20661]: cluster 2026-03-31T11:34:45.726076+0000 mgr.a (mgr.14152) 93 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:47.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:46 vm09 bash[20661]: audit 2026-03-31T11:34:46.619997+0000 mon.a (mon.0) 371 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:46 vm09 bash[20661]: audit 2026-03-31T11:34:46.619997+0000 mon.a (mon.0) 371 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:46 vm09 bash[20661]: audit 2026-03-31T11:34:46.622926+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:46 vm09 bash[20661]: audit 2026-03-31T11:34:46.622926+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:46 vm09 bash[20661]: audit 2026-03-31T11:34:46.626262+0000 mon.a (mon.0) 373 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:47.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:46 vm09 bash[20661]: audit 2026-03-31T11:34:46.626262+0000 mon.a (mon.0) 373 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:34:48.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:47 vm06 bash[20457]: cluster 2026-03-31T11:34:46.623249+0000 mgr.a (mgr.14152) 94 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:48.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:47 vm06 bash[20457]: cluster 2026-03-31T11:34:46.623249+0000 mgr.a (mgr.14152) 94 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:48.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:47 vm06 bash[20457]: cluster 2026-03-31T11:34:46.623333+0000 mgr.a (mgr.14152) 95 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:48.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:47 vm06 bash[20457]: cluster 2026-03-31T11:34:46.623333+0000 mgr.a (mgr.14152) 95 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:48.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:47 vm09 bash[20661]: cluster 2026-03-31T11:34:46.623249+0000 mgr.a (mgr.14152) 94 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:48.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:47 vm09 bash[20661]: cluster 2026-03-31T11:34:46.623249+0000 mgr.a (mgr.14152) 94 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:48.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:47 vm09 bash[20661]: cluster 2026-03-31T11:34:46.623333+0000 mgr.a (mgr.14152) 95 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:48.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:47 vm09 bash[20661]: cluster 2026-03-31T11:34:46.623333+0000 mgr.a (mgr.14152) 95 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:50.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:49 vm06 bash[20457]: cluster 2026-03-31T11:34:48.623579+0000 mgr.a (mgr.14152) 96 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:50.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:49 vm06 bash[20457]: cluster 2026-03-31T11:34:48.623579+0000 mgr.a (mgr.14152) 96 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:50.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:49 vm09 bash[20661]: cluster 2026-03-31T11:34:48.623579+0000 mgr.a (mgr.14152) 96 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:50.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:49 vm09 bash[20661]: cluster 2026-03-31T11:34:48.623579+0000 mgr.a (mgr.14152) 96 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:52.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:51 vm06 bash[20457]: cluster 2026-03-31T11:34:50.623802+0000 mgr.a (mgr.14152) 97 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:52.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:51 vm06 bash[20457]: cluster 2026-03-31T11:34:50.623802+0000 mgr.a (mgr.14152) 97 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:52.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:51 vm09 bash[20661]: cluster 2026-03-31T11:34:50.623802+0000 mgr.a (mgr.14152) 97 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:52.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:51 vm09 bash[20661]: cluster 2026-03-31T11:34:50.623802+0000 mgr.a (mgr.14152) 97 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:54.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:53 vm06 bash[20457]: cluster 2026-03-31T11:34:52.623939+0000 mgr.a (mgr.14152) 98 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:54.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:53 vm06 bash[20457]: cluster 2026-03-31T11:34:52.623939+0000 mgr.a (mgr.14152) 98 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:54.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:53 vm09 bash[20661]: cluster 2026-03-31T11:34:52.623939+0000 mgr.a (mgr.14152) 98 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:54.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:53 vm09 bash[20661]: cluster 2026-03-31T11:34:52.623939+0000 mgr.a (mgr.14152) 98 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:56.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:55 vm06 bash[20457]: cluster 2026-03-31T11:34:54.624108+0000 mgr.a (mgr.14152) 99 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:56.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:55 vm06 bash[20457]: cluster 2026-03-31T11:34:54.624108+0000 mgr.a (mgr.14152) 99 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:56.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:55 vm09 bash[20661]: cluster 2026-03-31T11:34:54.624108+0000 mgr.a (mgr.14152) 99 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:56.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:55 vm09 bash[20661]: cluster 2026-03-31T11:34:54.624108+0000 mgr.a (mgr.14152) 99 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:58.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:57 vm06 bash[20457]: cluster 2026-03-31T11:34:56.624283+0000 mgr.a (mgr.14152) 100 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:58.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:57 vm06 bash[20457]: cluster 2026-03-31T11:34:56.624283+0000 mgr.a (mgr.14152) 100 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:58.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:57 vm09 bash[20661]: cluster 2026-03-31T11:34:56.624283+0000 mgr.a (mgr.14152) 100 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:34:58.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:57 vm09 bash[20661]: cluster 2026-03-31T11:34:56.624283+0000 mgr.a (mgr.14152) 100 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:00.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:59 vm06 bash[20457]: cluster 2026-03-31T11:34:58.624478+0000 mgr.a (mgr.14152) 101 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:00.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:34:59 vm06 bash[20457]: cluster 2026-03-31T11:34:58.624478+0000 mgr.a (mgr.14152) 101 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:00.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:59 vm09 bash[20661]: cluster 2026-03-31T11:34:58.624478+0000 mgr.a (mgr.14152) 101 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:00.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:34:59 vm09 bash[20661]: cluster 2026-03-31T11:34:58.624478+0000 mgr.a (mgr.14152) 101 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:01.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:01 vm06 bash[20457]: cluster 2026-03-31T11:35:00.624676+0000 mgr.a (mgr.14152) 102 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:01.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:01 vm06 bash[20457]: cluster 2026-03-31T11:35:00.624676+0000 mgr.a (mgr.14152) 102 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:02.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:01 vm09 bash[20661]: cluster 2026-03-31T11:35:00.624676+0000 mgr.a (mgr.14152) 102 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:02.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:01 vm09 bash[20661]: cluster 2026-03-31T11:35:00.624676+0000 mgr.a (mgr.14152) 102 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: audit 2026-03-31T11:35:01.787285+0000 mon.a (mon.0) 374 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: audit 2026-03-31T11:35:01.787285+0000 mon.a (mon.0) 374 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: audit 2026-03-31T11:35:01.790258+0000 mon.a (mon.0) 375 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: audit 2026-03-31T11:35:01.790258+0000 mon.a (mon.0) 375 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: cluster 2026-03-31T11:35:01.790522+0000 mgr.a (mgr.14152) 103 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: cluster 2026-03-31T11:35:01.790522+0000 mgr.a (mgr.14152) 103 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: cluster 2026-03-31T11:35:01.790567+0000 mgr.a (mgr.14152) 104 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: cluster 2026-03-31T11:35:01.790567+0000 mgr.a (mgr.14152) 104 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: audit 2026-03-31T11:35:01.793363+0000 mon.a (mon.0) 376 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:02 vm06 bash[20457]: audit 2026-03-31T11:35:01.793363+0000 mon.a (mon.0) 376 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: audit 2026-03-31T11:35:01.787285+0000 mon.a (mon.0) 374 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: audit 2026-03-31T11:35:01.787285+0000 mon.a (mon.0) 374 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: audit 2026-03-31T11:35:01.790258+0000 mon.a (mon.0) 375 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: audit 2026-03-31T11:35:01.790258+0000 mon.a (mon.0) 375 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: cluster 2026-03-31T11:35:01.790522+0000 mgr.a (mgr.14152) 103 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: cluster 2026-03-31T11:35:01.790522+0000 mgr.a (mgr.14152) 103 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: cluster 2026-03-31T11:35:01.790567+0000 mgr.a (mgr.14152) 104 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: cluster 2026-03-31T11:35:01.790567+0000 mgr.a (mgr.14152) 104 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: audit 2026-03-31T11:35:01.793363+0000 mon.a (mon.0) 376 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:02 vm09 bash[20661]: audit 2026-03-31T11:35:01.793363+0000 mon.a (mon.0) 376 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:05.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:04 vm06 bash[20457]: cluster 2026-03-31T11:35:03.790807+0000 mgr.a (mgr.14152) 105 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:05.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:04 vm06 bash[20457]: cluster 2026-03-31T11:35:03.790807+0000 mgr.a (mgr.14152) 105 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:05.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:04 vm09 bash[20661]: cluster 2026-03-31T11:35:03.790807+0000 mgr.a (mgr.14152) 105 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:05.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:04 vm09 bash[20661]: cluster 2026-03-31T11:35:03.790807+0000 mgr.a (mgr.14152) 105 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:06 vm06 bash[20457]: cluster 2026-03-31T11:35:05.791036+0000 mgr.a (mgr.14152) 106 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:06 vm06 bash[20457]: cluster 2026-03-31T11:35:05.791036+0000 mgr.a (mgr.14152) 106 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:06 vm06 bash[20457]: audit 2026-03-31T11:35:06.683491+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:06 vm06 bash[20457]: audit 2026-03-31T11:35:06.683491+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:06 vm06 bash[20457]: audit 2026-03-31T11:35:06.687732+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:06 vm06 bash[20457]: audit 2026-03-31T11:35:06.687732+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:06 vm06 bash[20457]: audit 2026-03-31T11:35:06.690961+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:06 vm06 bash[20457]: audit 2026-03-31T11:35:06.690961+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:06 vm09 bash[20661]: cluster 2026-03-31T11:35:05.791036+0000 mgr.a (mgr.14152) 106 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:07.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:06 vm09 bash[20661]: cluster 2026-03-31T11:35:05.791036+0000 mgr.a (mgr.14152) 106 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:07.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:06 vm09 bash[20661]: audit 2026-03-31T11:35:06.683491+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:06 vm09 bash[20661]: audit 2026-03-31T11:35:06.683491+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:06 vm09 bash[20661]: audit 2026-03-31T11:35:06.687732+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:06 vm09 bash[20661]: audit 2026-03-31T11:35:06.687732+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:06 vm09 bash[20661]: audit 2026-03-31T11:35:06.690961+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:07.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:06 vm09 bash[20661]: audit 2026-03-31T11:35:06.690961+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:08.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:07 vm06 bash[20457]: cluster 2026-03-31T11:35:06.688015+0000 mgr.a (mgr.14152) 107 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:08.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:07 vm06 bash[20457]: cluster 2026-03-31T11:35:06.688015+0000 mgr.a (mgr.14152) 107 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:08.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:07 vm09 bash[20661]: cluster 2026-03-31T11:35:06.688015+0000 mgr.a (mgr.14152) 107 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:08.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:07 vm09 bash[20661]: cluster 2026-03-31T11:35:06.688015+0000 mgr.a (mgr.14152) 107 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:10.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:09 vm06 bash[20457]: cluster 2026-03-31T11:35:08.688245+0000 mgr.a (mgr.14152) 108 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:10.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:09 vm06 bash[20457]: cluster 2026-03-31T11:35:08.688245+0000 mgr.a (mgr.14152) 108 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:10.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:09 vm09 bash[20661]: cluster 2026-03-31T11:35:08.688245+0000 mgr.a (mgr.14152) 108 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:10.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:09 vm09 bash[20661]: cluster 2026-03-31T11:35:08.688245+0000 mgr.a (mgr.14152) 108 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:12.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:11 vm09 bash[20661]: cluster 2026-03-31T11:35:10.688442+0000 mgr.a (mgr.14152) 109 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:12.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:11 vm09 bash[20661]: cluster 2026-03-31T11:35:10.688442+0000 mgr.a (mgr.14152) 109 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:12.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:11 vm06 bash[20457]: cluster 2026-03-31T11:35:10.688442+0000 mgr.a (mgr.14152) 109 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:12.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:11 vm06 bash[20457]: cluster 2026-03-31T11:35:10.688442+0000 mgr.a (mgr.14152) 109 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:14.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:13 vm09 bash[20661]: cluster 2026-03-31T11:35:12.688614+0000 mgr.a (mgr.14152) 110 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:14.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:13 vm09 bash[20661]: cluster 2026-03-31T11:35:12.688614+0000 mgr.a (mgr.14152) 110 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:14.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:13 vm06 bash[20457]: cluster 2026-03-31T11:35:12.688614+0000 mgr.a (mgr.14152) 110 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:14.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:13 vm06 bash[20457]: cluster 2026-03-31T11:35:12.688614+0000 mgr.a (mgr.14152) 110 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:16.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:15 vm09 bash[20661]: cluster 2026-03-31T11:35:14.688816+0000 mgr.a (mgr.14152) 111 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:16.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:15 vm09 bash[20661]: cluster 2026-03-31T11:35:14.688816+0000 mgr.a (mgr.14152) 111 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:16.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:15 vm06 bash[20457]: cluster 2026-03-31T11:35:14.688816+0000 mgr.a (mgr.14152) 111 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:16.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:15 vm06 bash[20457]: cluster 2026-03-31T11:35:14.688816+0000 mgr.a (mgr.14152) 111 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:18.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:17 vm09 bash[20661]: cluster 2026-03-31T11:35:16.689112+0000 mgr.a (mgr.14152) 112 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:18.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:17 vm09 bash[20661]: cluster 2026-03-31T11:35:16.689112+0000 mgr.a (mgr.14152) 112 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:18.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:17 vm06 bash[20457]: cluster 2026-03-31T11:35:16.689112+0000 mgr.a (mgr.14152) 112 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:18.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:17 vm06 bash[20457]: cluster 2026-03-31T11:35:16.689112+0000 mgr.a (mgr.14152) 112 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:19.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:18 vm09 bash[20661]: cluster 2026-03-31T11:35:18.689319+0000 mgr.a (mgr.14152) 113 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:19.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:18 vm09 bash[20661]: cluster 2026-03-31T11:35:18.689319+0000 mgr.a (mgr.14152) 113 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:19.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:18 vm06 bash[20457]: cluster 2026-03-31T11:35:18.689319+0000 mgr.a (mgr.14152) 113 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:19.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:18 vm06 bash[20457]: cluster 2026-03-31T11:35:18.689319+0000 mgr.a (mgr.14152) 113 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:22.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:21 vm06 bash[20457]: cluster 2026-03-31T11:35:20.689574+0000 mgr.a (mgr.14152) 114 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:22.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:21 vm06 bash[20457]: cluster 2026-03-31T11:35:20.689574+0000 mgr.a (mgr.14152) 114 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:22.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:21 vm09 bash[20661]: cluster 2026-03-31T11:35:20.689574+0000 mgr.a (mgr.14152) 114 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:22.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:21 vm09 bash[20661]: cluster 2026-03-31T11:35:20.689574+0000 mgr.a (mgr.14152) 114 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:23.137 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:22 vm06 bash[20457]: audit 2026-03-31T11:35:21.850581+0000 mon.a (mon.0) 380 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.137 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:22 vm06 bash[20457]: audit 2026-03-31T11:35:21.850581+0000 mon.a (mon.0) 380 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.137 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:22 vm06 bash[20457]: audit 2026-03-31T11:35:21.853798+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.137 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:22 vm06 bash[20457]: audit 2026-03-31T11:35:21.853798+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.137 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:22 vm06 bash[20457]: cluster 2026-03-31T11:35:21.854155+0000 mgr.a (mgr.14152) 115 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:23.137 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:22 vm06 bash[20457]: cluster 2026-03-31T11:35:21.854155+0000 mgr.a (mgr.14152) 115 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:23.137 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:22 vm06 bash[20457]: audit 2026-03-31T11:35:21.857029+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.137 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:22 vm06 bash[20457]: audit 2026-03-31T11:35:21.857029+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:22 vm09 bash[20661]: audit 2026-03-31T11:35:21.850581+0000 mon.a (mon.0) 380 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:22 vm09 bash[20661]: audit 2026-03-31T11:35:21.850581+0000 mon.a (mon.0) 380 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:22 vm09 bash[20661]: audit 2026-03-31T11:35:21.853798+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:22 vm09 bash[20661]: audit 2026-03-31T11:35:21.853798+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:22 vm09 bash[20661]: cluster 2026-03-31T11:35:21.854155+0000 mgr.a (mgr.14152) 115 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:23.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:22 vm09 bash[20661]: cluster 2026-03-31T11:35:21.854155+0000 mgr.a (mgr.14152) 115 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:23.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:22 vm09 bash[20661]: audit 2026-03-31T11:35:21.857029+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:23.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:22 vm09 bash[20661]: audit 2026-03-31T11:35:21.857029+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:25.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:24 vm06 bash[20457]: cluster 2026-03-31T11:35:23.854541+0000 mgr.a (mgr.14152) 116 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:25.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:24 vm06 bash[20457]: cluster 2026-03-31T11:35:23.854541+0000 mgr.a (mgr.14152) 116 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:24 vm09 bash[20661]: cluster 2026-03-31T11:35:23.854541+0000 mgr.a (mgr.14152) 116 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:24 vm09 bash[20661]: cluster 2026-03-31T11:35:23.854541+0000 mgr.a (mgr.14152) 116 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: cluster 2026-03-31T11:35:25.854801+0000 mgr.a (mgr.14152) 117 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: cluster 2026-03-31T11:35:25.854801+0000 mgr.a (mgr.14152) 117 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: audit 2026-03-31T11:35:26.749811+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: audit 2026-03-31T11:35:26.749811+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: audit 2026-03-31T11:35:26.752866+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: audit 2026-03-31T11:35:26.752866+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: cluster 2026-03-31T11:35:26.753287+0000 mgr.a (mgr.14152) 118 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: cluster 2026-03-31T11:35:26.753287+0000 mgr.a (mgr.14152) 118 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: audit 2026-03-31T11:35:26.756112+0000 mon.a (mon.0) 385 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:27 vm06 bash[20457]: audit 2026-03-31T11:35:26.756112+0000 mon.a (mon.0) 385 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: cluster 2026-03-31T11:35:25.854801+0000 mgr.a (mgr.14152) 117 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: cluster 2026-03-31T11:35:25.854801+0000 mgr.a (mgr.14152) 117 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: audit 2026-03-31T11:35:26.749811+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: audit 2026-03-31T11:35:26.749811+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: audit 2026-03-31T11:35:26.752866+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: audit 2026-03-31T11:35:26.752866+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: cluster 2026-03-31T11:35:26.753287+0000 mgr.a (mgr.14152) 118 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: cluster 2026-03-31T11:35:26.753287+0000 mgr.a (mgr.14152) 118 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:28.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: audit 2026-03-31T11:35:26.756112+0000 mon.a (mon.0) 385 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:28.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:27 vm09 bash[20661]: audit 2026-03-31T11:35:26.756112+0000 mon.a (mon.0) 385 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:30.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:29 vm06 bash[20457]: cluster 2026-03-31T11:35:28.753469+0000 mgr.a (mgr.14152) 119 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:30.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:29 vm06 bash[20457]: cluster 2026-03-31T11:35:28.753469+0000 mgr.a (mgr.14152) 119 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:30.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:29 vm09 bash[20661]: cluster 2026-03-31T11:35:28.753469+0000 mgr.a (mgr.14152) 119 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:30.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:29 vm09 bash[20661]: cluster 2026-03-31T11:35:28.753469+0000 mgr.a (mgr.14152) 119 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:32.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:31 vm06 bash[20457]: cluster 2026-03-31T11:35:30.753721+0000 mgr.a (mgr.14152) 120 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:32.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:31 vm06 bash[20457]: cluster 2026-03-31T11:35:30.753721+0000 mgr.a (mgr.14152) 120 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:32.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:31 vm09 bash[20661]: cluster 2026-03-31T11:35:30.753721+0000 mgr.a (mgr.14152) 120 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:32.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:31 vm09 bash[20661]: cluster 2026-03-31T11:35:30.753721+0000 mgr.a (mgr.14152) 120 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:34.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:33 vm06 bash[20457]: cluster 2026-03-31T11:35:32.753971+0000 mgr.a (mgr.14152) 121 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:34.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:33 vm06 bash[20457]: cluster 2026-03-31T11:35:32.753971+0000 mgr.a (mgr.14152) 121 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:34.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:33 vm09 bash[20661]: cluster 2026-03-31T11:35:32.753971+0000 mgr.a (mgr.14152) 121 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:34.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:33 vm09 bash[20661]: cluster 2026-03-31T11:35:32.753971+0000 mgr.a (mgr.14152) 121 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:36.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:35 vm06 bash[20457]: cluster 2026-03-31T11:35:34.754222+0000 mgr.a (mgr.14152) 122 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:36.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:35 vm06 bash[20457]: cluster 2026-03-31T11:35:34.754222+0000 mgr.a (mgr.14152) 122 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:36.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:35 vm09 bash[20661]: cluster 2026-03-31T11:35:34.754222+0000 mgr.a (mgr.14152) 122 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:36.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:35 vm09 bash[20661]: cluster 2026-03-31T11:35:34.754222+0000 mgr.a (mgr.14152) 122 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:38.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:37 vm06 bash[20457]: cluster 2026-03-31T11:35:36.754482+0000 mgr.a (mgr.14152) 123 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:38.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:37 vm06 bash[20457]: cluster 2026-03-31T11:35:36.754482+0000 mgr.a (mgr.14152) 123 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:38.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:37 vm09 bash[20661]: cluster 2026-03-31T11:35:36.754482+0000 mgr.a (mgr.14152) 123 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:38.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:37 vm09 bash[20661]: cluster 2026-03-31T11:35:36.754482+0000 mgr.a (mgr.14152) 123 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:40.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:39 vm06 bash[20457]: cluster 2026-03-31T11:35:38.754672+0000 mgr.a (mgr.14152) 124 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:40.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:39 vm06 bash[20457]: cluster 2026-03-31T11:35:38.754672+0000 mgr.a (mgr.14152) 124 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:40.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:39 vm09 bash[20661]: cluster 2026-03-31T11:35:38.754672+0000 mgr.a (mgr.14152) 124 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:40.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:39 vm09 bash[20661]: cluster 2026-03-31T11:35:38.754672+0000 mgr.a (mgr.14152) 124 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:42.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:41 vm06 bash[20457]: cluster 2026-03-31T11:35:40.754916+0000 mgr.a (mgr.14152) 125 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:42.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:41 vm06 bash[20457]: cluster 2026-03-31T11:35:40.754916+0000 mgr.a (mgr.14152) 125 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:42.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:41 vm09 bash[20661]: cluster 2026-03-31T11:35:40.754916+0000 mgr.a (mgr.14152) 125 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:42.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:41 vm09 bash[20661]: cluster 2026-03-31T11:35:40.754916+0000 mgr.a (mgr.14152) 125 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:43.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:42 vm06 bash[20457]: audit 2026-03-31T11:35:41.914579+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:42 vm06 bash[20457]: audit 2026-03-31T11:35:41.914579+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:42 vm06 bash[20457]: audit 2026-03-31T11:35:41.917729+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:42 vm06 bash[20457]: audit 2026-03-31T11:35:41.917729+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:42 vm06 bash[20457]: cluster 2026-03-31T11:35:41.918037+0000 mgr.a (mgr.14152) 126 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:43.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:42 vm06 bash[20457]: cluster 2026-03-31T11:35:41.918037+0000 mgr.a (mgr.14152) 126 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:43.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:42 vm06 bash[20457]: audit 2026-03-31T11:35:41.920874+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:42 vm06 bash[20457]: audit 2026-03-31T11:35:41.920874+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:42 vm09 bash[20661]: audit 2026-03-31T11:35:41.914579+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:42 vm09 bash[20661]: audit 2026-03-31T11:35:41.914579+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:42 vm09 bash[20661]: audit 2026-03-31T11:35:41.917729+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:42 vm09 bash[20661]: audit 2026-03-31T11:35:41.917729+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:42 vm09 bash[20661]: cluster 2026-03-31T11:35:41.918037+0000 mgr.a (mgr.14152) 126 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:43.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:42 vm09 bash[20661]: cluster 2026-03-31T11:35:41.918037+0000 mgr.a (mgr.14152) 126 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:43.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:42 vm09 bash[20661]: audit 2026-03-31T11:35:41.920874+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:43.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:42 vm09 bash[20661]: audit 2026-03-31T11:35:41.920874+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:45.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:44 vm06 bash[20457]: cluster 2026-03-31T11:35:43.918306+0000 mgr.a (mgr.14152) 127 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:45.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:44 vm06 bash[20457]: cluster 2026-03-31T11:35:43.918306+0000 mgr.a (mgr.14152) 127 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:45.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:44 vm09 bash[20661]: cluster 2026-03-31T11:35:43.918306+0000 mgr.a (mgr.14152) 127 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:45.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:44 vm09 bash[20661]: cluster 2026-03-31T11:35:43.918306+0000 mgr.a (mgr.14152) 127 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: cluster 2026-03-31T11:35:45.918578+0000 mgr.a (mgr.14152) 128 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: cluster 2026-03-31T11:35:45.918578+0000 mgr.a (mgr.14152) 128 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: audit 2026-03-31T11:35:46.812391+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: audit 2026-03-31T11:35:46.812391+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: audit 2026-03-31T11:35:46.816433+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: audit 2026-03-31T11:35:46.816433+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: cluster 2026-03-31T11:35:46.816772+0000 mgr.a (mgr.14152) 129 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: cluster 2026-03-31T11:35:46.816772+0000 mgr.a (mgr.14152) 129 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: cluster 2026-03-31T11:35:46.816844+0000 mgr.a (mgr.14152) 130 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: cluster 2026-03-31T11:35:46.816844+0000 mgr.a (mgr.14152) 130 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: audit 2026-03-31T11:35:46.819793+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:47 vm06 bash[20457]: audit 2026-03-31T11:35:46.819793+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: cluster 2026-03-31T11:35:45.918578+0000 mgr.a (mgr.14152) 128 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: cluster 2026-03-31T11:35:45.918578+0000 mgr.a (mgr.14152) 128 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: audit 2026-03-31T11:35:46.812391+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: audit 2026-03-31T11:35:46.812391+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: audit 2026-03-31T11:35:46.816433+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: audit 2026-03-31T11:35:46.816433+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: cluster 2026-03-31T11:35:46.816772+0000 mgr.a (mgr.14152) 129 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: cluster 2026-03-31T11:35:46.816772+0000 mgr.a (mgr.14152) 129 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: cluster 2026-03-31T11:35:46.816844+0000 mgr.a (mgr.14152) 130 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: cluster 2026-03-31T11:35:46.816844+0000 mgr.a (mgr.14152) 130 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: audit 2026-03-31T11:35:46.819793+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:47 vm09 bash[20661]: audit 2026-03-31T11:35:46.819793+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:35:50.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:49 vm09 bash[20661]: cluster 2026-03-31T11:35:48.817024+0000 mgr.a (mgr.14152) 131 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:50.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:49 vm09 bash[20661]: cluster 2026-03-31T11:35:48.817024+0000 mgr.a (mgr.14152) 131 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:50.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:49 vm06 bash[20457]: cluster 2026-03-31T11:35:48.817024+0000 mgr.a (mgr.14152) 131 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:50.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:49 vm06 bash[20457]: cluster 2026-03-31T11:35:48.817024+0000 mgr.a (mgr.14152) 131 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:52.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:51 vm09 bash[20661]: cluster 2026-03-31T11:35:50.817248+0000 mgr.a (mgr.14152) 132 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:52.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:51 vm09 bash[20661]: cluster 2026-03-31T11:35:50.817248+0000 mgr.a (mgr.14152) 132 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:52.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:51 vm06 bash[20457]: cluster 2026-03-31T11:35:50.817248+0000 mgr.a (mgr.14152) 132 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:52.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:51 vm06 bash[20457]: cluster 2026-03-31T11:35:50.817248+0000 mgr.a (mgr.14152) 132 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:54.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:53 vm06 bash[20457]: cluster 2026-03-31T11:35:52.817491+0000 mgr.a (mgr.14152) 133 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:54.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:53 vm06 bash[20457]: cluster 2026-03-31T11:35:52.817491+0000 mgr.a (mgr.14152) 133 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:54.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:53 vm09 bash[20661]: cluster 2026-03-31T11:35:52.817491+0000 mgr.a (mgr.14152) 133 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:54.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:53 vm09 bash[20661]: cluster 2026-03-31T11:35:52.817491+0000 mgr.a (mgr.14152) 133 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:55.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:54 vm06 bash[20457]: cluster 2026-03-31T11:35:54.817688+0000 mgr.a (mgr.14152) 134 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:55.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:54 vm06 bash[20457]: cluster 2026-03-31T11:35:54.817688+0000 mgr.a (mgr.14152) 134 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:55.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:54 vm09 bash[20661]: cluster 2026-03-31T11:35:54.817688+0000 mgr.a (mgr.14152) 134 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:55.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:54 vm09 bash[20661]: cluster 2026-03-31T11:35:54.817688+0000 mgr.a (mgr.14152) 134 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:57.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:56 vm06 bash[20457]: cluster 2026-03-31T11:35:56.817939+0000 mgr.a (mgr.14152) 135 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:57.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:56 vm06 bash[20457]: cluster 2026-03-31T11:35:56.817939+0000 mgr.a (mgr.14152) 135 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:57.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:56 vm09 bash[20661]: cluster 2026-03-31T11:35:56.817939+0000 mgr.a (mgr.14152) 135 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:57.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:56 vm09 bash[20661]: cluster 2026-03-31T11:35:56.817939+0000 mgr.a (mgr.14152) 135 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:59.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:58 vm06 bash[20457]: cluster 2026-03-31T11:35:58.818219+0000 mgr.a (mgr.14152) 136 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:59.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:35:58 vm06 bash[20457]: cluster 2026-03-31T11:35:58.818219+0000 mgr.a (mgr.14152) 136 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:59.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:58 vm09 bash[20661]: cluster 2026-03-31T11:35:58.818219+0000 mgr.a (mgr.14152) 136 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:35:59.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:35:58 vm09 bash[20661]: cluster 2026-03-31T11:35:58.818219+0000 mgr.a (mgr.14152) 136 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:01.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:00 vm06 bash[20457]: cluster 2026-03-31T11:36:00.818461+0000 mgr.a (mgr.14152) 137 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:01.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:00 vm06 bash[20457]: cluster 2026-03-31T11:36:00.818461+0000 mgr.a (mgr.14152) 137 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:01.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:00 vm09 bash[20661]: cluster 2026-03-31T11:36:00.818461+0000 mgr.a (mgr.14152) 137 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:01.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:00 vm09 bash[20661]: cluster 2026-03-31T11:36:00.818461+0000 mgr.a (mgr.14152) 137 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: audit 2026-03-31T11:36:01.978929+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: audit 2026-03-31T11:36:01.978929+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: audit 2026-03-31T11:36:01.982647+0000 mon.a (mon.0) 393 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: audit 2026-03-31T11:36:01.982647+0000 mon.a (mon.0) 393 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: cluster 2026-03-31T11:36:01.982971+0000 mgr.a (mgr.14152) 138 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: cluster 2026-03-31T11:36:01.982971+0000 mgr.a (mgr.14152) 138 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: cluster 2026-03-31T11:36:01.983105+0000 mgr.a (mgr.14152) 139 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: cluster 2026-03-31T11:36:01.983105+0000 mgr.a (mgr.14152) 139 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: audit 2026-03-31T11:36:01.986213+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:02 vm06 bash[20457]: audit 2026-03-31T11:36:01.986213+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: audit 2026-03-31T11:36:01.978929+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: audit 2026-03-31T11:36:01.978929+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: audit 2026-03-31T11:36:01.982647+0000 mon.a (mon.0) 393 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: audit 2026-03-31T11:36:01.982647+0000 mon.a (mon.0) 393 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: cluster 2026-03-31T11:36:01.982971+0000 mgr.a (mgr.14152) 138 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: cluster 2026-03-31T11:36:01.982971+0000 mgr.a (mgr.14152) 138 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: cluster 2026-03-31T11:36:01.983105+0000 mgr.a (mgr.14152) 139 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: cluster 2026-03-31T11:36:01.983105+0000 mgr.a (mgr.14152) 139 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: audit 2026-03-31T11:36:01.986213+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:03.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:02 vm09 bash[20661]: audit 2026-03-31T11:36:01.986213+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:05.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:04 vm06 bash[20457]: cluster 2026-03-31T11:36:03.983342+0000 mgr.a (mgr.14152) 140 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:05.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:04 vm06 bash[20457]: cluster 2026-03-31T11:36:03.983342+0000 mgr.a (mgr.14152) 140 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:05.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:04 vm09 bash[20661]: cluster 2026-03-31T11:36:03.983342+0000 mgr.a (mgr.14152) 140 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:05.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:04 vm09 bash[20661]: cluster 2026-03-31T11:36:03.983342+0000 mgr.a (mgr.14152) 140 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:07.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:06 vm06 bash[20457]: cluster 2026-03-31T11:36:05.983602+0000 mgr.a (mgr.14152) 141 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:07.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:06 vm06 bash[20457]: cluster 2026-03-31T11:36:05.983602+0000 mgr.a (mgr.14152) 141 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:07.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:06 vm06 bash[20457]: audit 2026-03-31T11:36:06.877498+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:06 vm06 bash[20457]: audit 2026-03-31T11:36:06.877498+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:06 vm06 bash[20457]: audit 2026-03-31T11:36:06.880641+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:06 vm06 bash[20457]: audit 2026-03-31T11:36:06.880641+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:06 vm06 bash[20457]: audit 2026-03-31T11:36:06.884002+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:06 vm06 bash[20457]: audit 2026-03-31T11:36:06.884002+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:06 vm09 bash[20661]: cluster 2026-03-31T11:36:05.983602+0000 mgr.a (mgr.14152) 141 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:07.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:06 vm09 bash[20661]: cluster 2026-03-31T11:36:05.983602+0000 mgr.a (mgr.14152) 141 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:07.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:06 vm09 bash[20661]: audit 2026-03-31T11:36:06.877498+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:06 vm09 bash[20661]: audit 2026-03-31T11:36:06.877498+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:06 vm09 bash[20661]: audit 2026-03-31T11:36:06.880641+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:06 vm09 bash[20661]: audit 2026-03-31T11:36:06.880641+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:06 vm09 bash[20661]: audit 2026-03-31T11:36:06.884002+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:07.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:06 vm09 bash[20661]: audit 2026-03-31T11:36:06.884002+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:08.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:07 vm06 bash[20457]: cluster 2026-03-31T11:36:06.880914+0000 mgr.a (mgr.14152) 142 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:08.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:07 vm06 bash[20457]: cluster 2026-03-31T11:36:06.880914+0000 mgr.a (mgr.14152) 142 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:08.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:07 vm09 bash[20661]: cluster 2026-03-31T11:36:06.880914+0000 mgr.a (mgr.14152) 142 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:08.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:07 vm09 bash[20661]: cluster 2026-03-31T11:36:06.880914+0000 mgr.a (mgr.14152) 142 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:10.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:10 vm06 bash[20457]: cluster 2026-03-31T11:36:08.881088+0000 mgr.a (mgr.14152) 143 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:10.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:10 vm06 bash[20457]: cluster 2026-03-31T11:36:08.881088+0000 mgr.a (mgr.14152) 143 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:10.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:10 vm09 bash[20661]: cluster 2026-03-31T11:36:08.881088+0000 mgr.a (mgr.14152) 143 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:10.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:10 vm09 bash[20661]: cluster 2026-03-31T11:36:08.881088+0000 mgr.a (mgr.14152) 143 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:11.101 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:10 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:11.101 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:36:10 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:12.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:12 vm06 bash[20457]: cluster 2026-03-31T11:36:10.881352+0000 mgr.a (mgr.14152) 144 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:12.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:12 vm06 bash[20457]: cluster 2026-03-31T11:36:10.881352+0000 mgr.a (mgr.14152) 144 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:12.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:12 vm06 bash[20457]: audit 2026-03-31T11:36:11.048461+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:12.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:12 vm06 bash[20457]: audit 2026-03-31T11:36:11.048461+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:12.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:12 vm06 bash[20457]: audit 2026-03-31T11:36:11.052780+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:12.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:12 vm06 bash[20457]: audit 2026-03-31T11:36:11.052780+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:12.317 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:12 vm09 bash[20661]: cluster 2026-03-31T11:36:10.881352+0000 mgr.a (mgr.14152) 144 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:12.317 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:12 vm09 bash[20661]: cluster 2026-03-31T11:36:10.881352+0000 mgr.a (mgr.14152) 144 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:12.317 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:12 vm09 bash[20661]: audit 2026-03-31T11:36:11.048461+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:12.317 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:12 vm09 bash[20661]: audit 2026-03-31T11:36:11.048461+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:12.317 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:12 vm09 bash[20661]: audit 2026-03-31T11:36:11.052780+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:12.317 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:12 vm09 bash[20661]: audit 2026-03-31T11:36:11.052780+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:12.879 INFO:teuthology.orchestra.run.vm09.stdout:Created osd(s) 1 on host 'vm09' 2026-03-31T11:36:12.954 DEBUG:teuthology.orchestra.run.vm09:osd.1> sudo journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@osd.1.service 2026-03-31T11:36:12.955 INFO:tasks.cephadm:Waiting for 2 OSDs to come up... 2026-03-31T11:36:12.955 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd stat -f json 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.482204+0000 mon.a (mon.0) 400 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.482204+0000 mon.a (mon.0) 400 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.482231+0000 mon.a (mon.0) 401 : cluster [INF] osd.1 failed (root=default,host=vm09) (connection refused reported by osd.0) 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.482231+0000 mon.a (mon.0) 401 : cluster [INF] osd.1 failed (root=default,host=vm09) (connection refused reported by osd.0) 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.682731+0000 mon.a (mon.0) 402 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.682731+0000 mon.a (mon.0) 402 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.682758+0000 mon.a (mon.0) 403 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.682758+0000 mon.a (mon.0) 403 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.682806+0000 mon.a (mon.0) 404 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: cluster 2026-03-31T11:36:12.682806+0000 mon.a (mon.0) 404 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: audit 2026-03-31T11:36:12.873237+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: audit 2026-03-31T11:36:12.873237+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: audit 2026-03-31T11:36:12.876594+0000 mon.a (mon.0) 406 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:13.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:13 vm09 bash[20661]: audit 2026-03-31T11:36:12.876594+0000 mon.a (mon.0) 406 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:13.197 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:36:13 vm09 bash[27160]: debug 2026-03-31T11:36:13.190+0000 7f03c37888c0 -1 osd.1 13 log_to_monitors true 2026-03-31T11:36:13.209 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.482204+0000 mon.a (mon.0) 400 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.482204+0000 mon.a (mon.0) 400 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.482231+0000 mon.a (mon.0) 401 : cluster [INF] osd.1 failed (root=default,host=vm09) (connection refused reported by osd.0) 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.482231+0000 mon.a (mon.0) 401 : cluster [INF] osd.1 failed (root=default,host=vm09) (connection refused reported by osd.0) 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.682731+0000 mon.a (mon.0) 402 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.682731+0000 mon.a (mon.0) 402 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.682758+0000 mon.a (mon.0) 403 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.682758+0000 mon.a (mon.0) 403 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.682806+0000 mon.a (mon.0) 404 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: cluster 2026-03-31T11:36:12.682806+0000 mon.a (mon.0) 404 : cluster [DBG] osd.1 reported immediately failed by osd.0 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: audit 2026-03-31T11:36:12.873237+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: audit 2026-03-31T11:36:12.873237+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: audit 2026-03-31T11:36:12.876594+0000 mon.a (mon.0) 406 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:13.226 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:13 vm06 bash[20457]: audit 2026-03-31T11:36:12.876594+0000 mon.a (mon.0) 406 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:13.509 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:13.560 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":14,"num_osds":2,"num_up_osds":1,"osd_up_since":1774956973,"num_in_osds":2,"osd_in_since":1774956846,"num_remapped_pgs":0} 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: cluster 2026-03-31T11:36:12.881568+0000 mgr.a (mgr.14152) 145 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: cluster 2026-03-31T11:36:12.881568+0000 mgr.a (mgr.14152) 145 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: cluster 2026-03-31T11:36:13.050457+0000 mon.a (mon.0) 407 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: cluster 2026-03-31T11:36:13.050457+0000 mon.a (mon.0) 407 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: cluster 2026-03-31T11:36:13.050509+0000 mon.a (mon.0) 408 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: cluster 2026-03-31T11:36:13.050509+0000 mon.a (mon.0) 408 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: cluster 2026-03-31T11:36:13.061311+0000 mon.a (mon.0) 409 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: cluster 2026-03-31T11:36:13.061311+0000 mon.a (mon.0) 409 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: audit 2026-03-31T11:36:13.201170+0000 mon.a (mon.0) 410 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: audit 2026-03-31T11:36:13.201170+0000 mon.a (mon.0) 410 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: audit 2026-03-31T11:36:13.509775+0000 mon.a (mon.0) 411 : audit [DBG] from='client.? 192.168.123.106:0/1135139712' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-31T11:36:14.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:14 vm06 bash[20457]: audit 2026-03-31T11:36:13.509775+0000 mon.a (mon.0) 411 : audit [DBG] from='client.? 192.168.123.106:0/1135139712' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: cluster 2026-03-31T11:36:12.881568+0000 mgr.a (mgr.14152) 145 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: cluster 2026-03-31T11:36:12.881568+0000 mgr.a (mgr.14152) 145 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: cluster 2026-03-31T11:36:13.050457+0000 mon.a (mon.0) 407 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: cluster 2026-03-31T11:36:13.050457+0000 mon.a (mon.0) 407 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: cluster 2026-03-31T11:36:13.050509+0000 mon.a (mon.0) 408 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: cluster 2026-03-31T11:36:13.050509+0000 mon.a (mon.0) 408 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: cluster 2026-03-31T11:36:13.061311+0000 mon.a (mon.0) 409 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: cluster 2026-03-31T11:36:13.061311+0000 mon.a (mon.0) 409 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: audit 2026-03-31T11:36:13.201170+0000 mon.a (mon.0) 410 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: audit 2026-03-31T11:36:13.201170+0000 mon.a (mon.0) 410 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: audit 2026-03-31T11:36:13.509775+0000 mon.a (mon.0) 411 : audit [DBG] from='client.? 192.168.123.106:0/1135139712' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-31T11:36:14.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:14 vm09 bash[20661]: audit 2026-03-31T11:36:13.509775+0000 mon.a (mon.0) 411 : audit [DBG] from='client.? 192.168.123.106:0/1135139712' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-31T11:36:14.335 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:36:14 vm09 bash[27160]: debug 2026-03-31T11:36:14.082+0000 7f03ba4d9640 -1 osd.1 13 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-31T11:36:14.561 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd stat -f json 2026-03-31T11:36:14.818 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:15.112 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.060843+0000 mon.a (mon.0) 412 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.060843+0000 mon.a (mon.0) 412 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: cluster 2026-03-31T11:36:14.062857+0000 mon.a (mon.0) 413 : cluster [DBG] osdmap e15: 2 total, 1 up, 2 in 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: cluster 2026-03-31T11:36:14.062857+0000 mon.a (mon.0) 413 : cluster [DBG] osdmap e15: 2 total, 1 up, 2 in 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.063189+0000 mon.a (mon.0) 414 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.063189+0000 mon.a (mon.0) 414 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.129648+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.129648+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.133367+0000 mon.a (mon.0) 416 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.133367+0000 mon.a (mon.0) 416 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.144188+0000 mon.a (mon.0) 417 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.144188+0000 mon.a (mon.0) 417 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.145329+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-31T11:36:15.123 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.145329+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: cephadm 2026-03-31T11:36:14.145630+0000 mgr.a (mgr.14152) 146 : cephadm [INF] Adjusting osd_memory_target on vm09 to 814.9M 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: cephadm 2026-03-31T11:36:14.145630+0000 mgr.a (mgr.14152) 146 : cephadm [INF] Adjusting osd_memory_target on vm09 to 814.9M 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: cephadm 2026-03-31T11:36:14.145998+0000 mgr.a (mgr.14152) 147 : cephadm [WRN] Unable to set osd_memory_target on vm09 to 854509977: error parsing value: Value '854509977' is below minimum 939524096 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: cephadm 2026-03-31T11:36:14.145998+0000 mgr.a (mgr.14152) 147 : cephadm [WRN] Unable to set osd_memory_target on vm09 to 854509977: error parsing value: Value '854509977' is below minimum 939524096 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.146280+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.146280+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.146645+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.146645+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: cluster 2026-03-31T11:36:14.147265+0000 mgr.a (mgr.14152) 148 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: cluster 2026-03-31T11:36:14.147265+0000 mgr.a (mgr.14152) 148 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.150491+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.150491+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.151742+0000 mon.a (mon.0) 422 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.151742+0000 mon.a (mon.0) 422 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.152724+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.152724+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.153109+0000 mon.a (mon.0) 424 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:15.124 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:15 vm06 bash[20457]: audit 2026-03-31T11:36:14.153109+0000 mon.a (mon.0) 424 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:15.191 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":16,"num_osds":2,"num_up_osds":2,"osd_up_since":1774956975,"num_in_osds":2,"osd_in_since":1774956846,"num_remapped_pgs":0} 2026-03-31T11:36:15.192 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd dump --format=json 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.060843+0000 mon.a (mon.0) 412 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.060843+0000 mon.a (mon.0) 412 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: cluster 2026-03-31T11:36:14.062857+0000 mon.a (mon.0) 413 : cluster [DBG] osdmap e15: 2 total, 1 up, 2 in 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: cluster 2026-03-31T11:36:14.062857+0000 mon.a (mon.0) 413 : cluster [DBG] osdmap e15: 2 total, 1 up, 2 in 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.063189+0000 mon.a (mon.0) 414 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.063189+0000 mon.a (mon.0) 414 : audit [INF] from='osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm09", "root=default"]} : dispatch 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.129648+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.129648+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.133367+0000 mon.a (mon.0) 416 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.133367+0000 mon.a (mon.0) 416 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.329 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.144188+0000 mon.a (mon.0) 417 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.144188+0000 mon.a (mon.0) 417 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.145329+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.145329+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: cephadm 2026-03-31T11:36:14.145630+0000 mgr.a (mgr.14152) 146 : cephadm [INF] Adjusting osd_memory_target on vm09 to 814.9M 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: cephadm 2026-03-31T11:36:14.145630+0000 mgr.a (mgr.14152) 146 : cephadm [INF] Adjusting osd_memory_target on vm09 to 814.9M 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: cephadm 2026-03-31T11:36:14.145998+0000 mgr.a (mgr.14152) 147 : cephadm [WRN] Unable to set osd_memory_target on vm09 to 854509977: error parsing value: Value '854509977' is below minimum 939524096 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: cephadm 2026-03-31T11:36:14.145998+0000 mgr.a (mgr.14152) 147 : cephadm [WRN] Unable to set osd_memory_target on vm09 to 854509977: error parsing value: Value '854509977' is below minimum 939524096 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.146280+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.146280+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.146645+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.146645+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: cluster 2026-03-31T11:36:14.147265+0000 mgr.a (mgr.14152) 148 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: cluster 2026-03-31T11:36:14.147265+0000 mgr.a (mgr.14152) 148 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.150491+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.150491+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.151742+0000 mon.a (mon.0) 422 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.151742+0000 mon.a (mon.0) 422 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.152724+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.152724+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.153109+0000 mon.a (mon.0) 424 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:15.330 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:15 vm09 bash[20661]: audit 2026-03-31T11:36:14.153109+0000 mon.a (mon.0) 424 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:15.437 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:15.745 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:15.745 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":16,"fsid":"48134446-2cf5-11f1-830d-4121aa8d72d5","created":"2026-03-31T11:33:11.294091+0000","modified":"2026-03-31T11:36:15.060853+0000","last_up_change":"2026-03-31T11:36:15.060853+0000","last_in_change":"2026-03-31T11:34:06.776864+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":7,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"88f60178-5fdf-4f09-a5f8-805f65d0338d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6803","nonce":3317286751}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6805","nonce":3317286751}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6809","nonce":3317286751}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6807","nonce":3317286751}]},"public_addr":"192.168.123.106:6803/3317286751","cluster_addr":"192.168.123.106:6805/3317286751","heartbeat_back_addr":"192.168.123.106:6809/3317286751","heartbeat_front_addr":"192.168.123.106:6807/3317286751","state":["exists","up"]},{"osd":1,"uuid":"dabbc60e-788e-440b-9f82-29039e83287a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":12,"last_clean_end":13,"up_from":16,"up_thru":0,"down_at":14,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6801","nonce":2178563013}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6803","nonce":2178563013}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6807","nonce":2178563013}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6805","nonce":2178563013}]},"public_addr":"192.168.123.109:6801/2178563013","cluster_addr":"192.168.123.109:6803/2178563013","heartbeat_back_addr":"192.168.123.109:6807/2178563013","heartbeat_front_addr":"192.168.123.109:6805/2178563013","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-03-31T11:34:05.374891+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-03-31T11:34:12.966607+0000","dead_epoch":14}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.106:0/1709057529":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/1073351636":"2026-04-01T11:33:34.349039+0000","192.168.123.106:6801/947852417":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/1554541810":"2026-04-01T11:33:22.475137+0000","192.168.123.106:0/708176341":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/2978357861":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6801/942312008":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6800/947852417":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/3320458621":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6800/942312008":"2026-04-01T11:33:22.475137+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-31T11:36:15.795 INFO:tasks.cephadm.ceph_manager.ceph:[] 2026-03-31T11:36:15.795 INFO:tasks.cephadm:Setting up client nodes... 2026-03-31T11:36:15.795 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph config log 1 --format=json 2026-03-31T11:36:16.034 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.060967+0000 mon.a (mon.0) 425 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.060967+0000 mon.a (mon.0) 425 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.060980+0000 mon.a (mon.0) 426 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.060980+0000 mon.a (mon.0) 426 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.065703+0000 mon.a (mon.0) 427 : cluster [INF] osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013] boot 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.065703+0000 mon.a (mon.0) 427 : cluster [INF] osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013] boot 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.065732+0000 mon.a (mon.0) 428 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.065732+0000 mon.a (mon.0) 428 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: audit 2026-03-31T11:36:15.066761+0000 mon.a (mon.0) 429 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: audit 2026-03-31T11:36:15.066761+0000 mon.a (mon.0) 429 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: audit 2026-03-31T11:36:15.112115+0000 mon.a (mon.0) 430 : audit [DBG] from='client.? 192.168.123.106:0/793539501' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: audit 2026-03-31T11:36:15.112115+0000 mon.a (mon.0) 430 : audit [DBG] from='client.? 192.168.123.106:0/793539501' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.132372+0000 mon.a (mon.0) 431 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.132372+0000 mon.a (mon.0) 431 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.132396+0000 mon.a (mon.0) 432 : cluster [INF] Cluster is now healthy 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: cluster 2026-03-31T11:36:15.132396+0000 mon.a (mon.0) 432 : cluster [INF] Cluster is now healthy 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: audit 2026-03-31T11:36:15.745313+0000 mon.a (mon.0) 433 : audit [DBG] from='client.? 192.168.123.106:0/1841900054' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:16.219 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:16 vm09 bash[20661]: audit 2026-03-31T11:36:15.745313+0000 mon.a (mon.0) 433 : audit [DBG] from='client.? 192.168.123.106:0/1841900054' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:16.318 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.060967+0000 mon.a (mon.0) 425 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.060967+0000 mon.a (mon.0) 425 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.060980+0000 mon.a (mon.0) 426 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.060980+0000 mon.a (mon.0) 426 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.065703+0000 mon.a (mon.0) 427 : cluster [INF] osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013] boot 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.065703+0000 mon.a (mon.0) 427 : cluster [INF] osd.1 [v2:192.168.123.109:6800/2178563013,v1:192.168.123.109:6801/2178563013] boot 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.065732+0000 mon.a (mon.0) 428 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.065732+0000 mon.a (mon.0) 428 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: audit 2026-03-31T11:36:15.066761+0000 mon.a (mon.0) 429 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: audit 2026-03-31T11:36:15.066761+0000 mon.a (mon.0) 429 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: audit 2026-03-31T11:36:15.112115+0000 mon.a (mon.0) 430 : audit [DBG] from='client.? 192.168.123.106:0/793539501' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: audit 2026-03-31T11:36:15.112115+0000 mon.a (mon.0) 430 : audit [DBG] from='client.? 192.168.123.106:0/793539501' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.132372+0000 mon.a (mon.0) 431 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.132372+0000 mon.a (mon.0) 431 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.132396+0000 mon.a (mon.0) 432 : cluster [INF] Cluster is now healthy 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: cluster 2026-03-31T11:36:15.132396+0000 mon.a (mon.0) 432 : cluster [INF] Cluster is now healthy 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: audit 2026-03-31T11:36:15.745313+0000 mon.a (mon.0) 433 : audit [DBG] from='client.? 192.168.123.106:0/1841900054' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:16.328 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:16 vm06 bash[20457]: audit 2026-03-31T11:36:15.745313+0000 mon.a (mon.0) 433 : audit [DBG] from='client.? 192.168.123.106:0/1841900054' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:16.368 INFO:teuthology.orchestra.run.vm06.stdout:[{"version":10,"timestamp":"2026-03-31T11:33:39.106407+0000","name":"","changes":[{"name":"mgr/mgr/cephadm/allow_ptrace","new_value":"true"}]}] 2026-03-31T11:36:16.368 INFO:tasks.ceph_manager:config epoch is 10 2026-03-31T11:36:16.368 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-31T11:36:16.368 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-31T11:36:16.368 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph mgr dump --format=json 2026-03-31T11:36:16.604 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:16.932 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:16.986 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":13,"flags":0,"active_gid":14152,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":197146321},{"type":"v1","addr":"192.168.123.106:6801","nonce":197146321}]},"active_addr":"192.168.123.106:6801/197146321","active_change":"2026-03-31T11:33:34.349149+0000","active_mgr_features":4541880224203014143,"available":true,"standbys":[{"gid":14208,"name":"b","mgr_features":4541880224203014143,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.3.1","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROM_ALERT_CREDENTIAL_CACHE_TTL":{"name":"PROM_ALERT_CREDENTIAL_CACHE_TTL","type":"int","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crypto_caller":{"name":"crypto_caller","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"prometheus_tls_secret_name":{"name":"prometheus_tls_secret_name","type":"str","level":"advanced","flags":0,"default_value":"rook-ceph-prometheus-server-tls","min":"","max":"","enum_allowed":[],"desc":"name of tls secret in k8s for prometheus","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"smb","can_run":true,"error_string":"","module_options":{"internal_store_backend":{"name":"internal_store_backend","type":"str","level":"dev","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"set internal store backend. for develoment and testing only","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_orchestration":{"name":"update_orchestration","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically update orchestration when smb resources are changed","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.3.1","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROM_ALERT_CREDENTIAL_CACHE_TTL":{"name":"PROM_ALERT_CREDENTIAL_CACHE_TTL","type":"int","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crypto_caller":{"name":"crypto_caller","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"prometheus_tls_secret_name":{"name":"prometheus_tls_secret_name","type":"str","level":"advanced","flags":0,"default_value":"rook-ceph-prometheus-server-tls","min":"","max":"","enum_allowed":[],"desc":"name of tls secret in k8s for prometheus","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"smb","can_run":true,"error_string":"","module_options":{"internal_store_backend":{"name":"internal_store_backend","type":"str","level":"dev","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"set internal store backend. for develoment and testing only","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_orchestration":{"name":"update_orchestration","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically update orchestration when smb resources are changed","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.106:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"tentacle":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":402153024}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":2293408541}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":2125631346}]}]} 2026-03-31T11:36:16.988 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-31T11:36:16.988 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-31T11:36:16.988 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd dump --format=json 2026-03-31T11:36:17.226 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:17.239 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:17 vm06 bash[20457]: cluster 2026-03-31T11:36:16.147502+0000 mgr.a (mgr.14152) 149 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:17.239 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:17 vm06 bash[20457]: cluster 2026-03-31T11:36:16.147502+0000 mgr.a (mgr.14152) 149 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:17.239 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:17 vm06 bash[20457]: audit 2026-03-31T11:36:16.318111+0000 mon.a (mon.0) 434 : audit [DBG] from='client.? 192.168.123.106:0/2976488068' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-31T11:36:17.239 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:17 vm06 bash[20457]: audit 2026-03-31T11:36:16.318111+0000 mon.a (mon.0) 434 : audit [DBG] from='client.? 192.168.123.106:0/2976488068' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-31T11:36:17.239 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:17 vm06 bash[20457]: audit 2026-03-31T11:36:16.930705+0000 mon.a (mon.0) 435 : audit [DBG] from='client.? 192.168.123.106:0/3745024783' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-31T11:36:17.239 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:17 vm06 bash[20457]: audit 2026-03-31T11:36:16.930705+0000 mon.a (mon.0) 435 : audit [DBG] from='client.? 192.168.123.106:0/3745024783' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-31T11:36:17.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:17 vm09 bash[20661]: cluster 2026-03-31T11:36:16.147502+0000 mgr.a (mgr.14152) 149 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:17.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:17 vm09 bash[20661]: cluster 2026-03-31T11:36:16.147502+0000 mgr.a (mgr.14152) 149 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:17.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:17 vm09 bash[20661]: audit 2026-03-31T11:36:16.318111+0000 mon.a (mon.0) 434 : audit [DBG] from='client.? 192.168.123.106:0/2976488068' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-31T11:36:17.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:17 vm09 bash[20661]: audit 2026-03-31T11:36:16.318111+0000 mon.a (mon.0) 434 : audit [DBG] from='client.? 192.168.123.106:0/2976488068' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-31T11:36:17.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:17 vm09 bash[20661]: audit 2026-03-31T11:36:16.930705+0000 mon.a (mon.0) 435 : audit [DBG] from='client.? 192.168.123.106:0/3745024783' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-31T11:36:17.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:17 vm09 bash[20661]: audit 2026-03-31T11:36:16.930705+0000 mon.a (mon.0) 435 : audit [DBG] from='client.? 192.168.123.106:0/3745024783' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-31T11:36:17.512 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:17.512 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":16,"fsid":"48134446-2cf5-11f1-830d-4121aa8d72d5","created":"2026-03-31T11:33:11.294091+0000","modified":"2026-03-31T11:36:15.060853+0000","last_up_change":"2026-03-31T11:36:15.060853+0000","last_in_change":"2026-03-31T11:34:06.776864+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":7,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"88f60178-5fdf-4f09-a5f8-805f65d0338d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6803","nonce":3317286751}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6805","nonce":3317286751}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6809","nonce":3317286751}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6807","nonce":3317286751}]},"public_addr":"192.168.123.106:6803/3317286751","cluster_addr":"192.168.123.106:6805/3317286751","heartbeat_back_addr":"192.168.123.106:6809/3317286751","heartbeat_front_addr":"192.168.123.106:6807/3317286751","state":["exists","up"]},{"osd":1,"uuid":"dabbc60e-788e-440b-9f82-29039e83287a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":12,"last_clean_end":13,"up_from":16,"up_thru":0,"down_at":14,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6801","nonce":2178563013}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6803","nonce":2178563013}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6807","nonce":2178563013}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6805","nonce":2178563013}]},"public_addr":"192.168.123.109:6801/2178563013","cluster_addr":"192.168.123.109:6803/2178563013","heartbeat_back_addr":"192.168.123.109:6807/2178563013","heartbeat_front_addr":"192.168.123.109:6805/2178563013","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-03-31T11:34:05.374891+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-03-31T11:34:12.966607+0000","dead_epoch":14}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.106:0/1709057529":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/1073351636":"2026-04-01T11:33:34.349039+0000","192.168.123.106:6801/947852417":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/1554541810":"2026-04-01T11:33:22.475137+0000","192.168.123.106:0/708176341":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/2978357861":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6801/942312008":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6800/947852417":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/3320458621":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6800/942312008":"2026-04-01T11:33:22.475137+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-31T11:36:17.559 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-31T11:36:17.559 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd dump --format=json 2026-03-31T11:36:17.789 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:18.086 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:18.086 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":16,"fsid":"48134446-2cf5-11f1-830d-4121aa8d72d5","created":"2026-03-31T11:33:11.294091+0000","modified":"2026-03-31T11:36:15.060853+0000","last_up_change":"2026-03-31T11:36:15.060853+0000","last_in_change":"2026-03-31T11:34:06.776864+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":7,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"88f60178-5fdf-4f09-a5f8-805f65d0338d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6803","nonce":3317286751}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6805","nonce":3317286751}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6809","nonce":3317286751}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3317286751},{"type":"v1","addr":"192.168.123.106:6807","nonce":3317286751}]},"public_addr":"192.168.123.106:6803/3317286751","cluster_addr":"192.168.123.106:6805/3317286751","heartbeat_back_addr":"192.168.123.106:6809/3317286751","heartbeat_front_addr":"192.168.123.106:6807/3317286751","state":["exists","up"]},{"osd":1,"uuid":"dabbc60e-788e-440b-9f82-29039e83287a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":12,"last_clean_end":13,"up_from":16,"up_thru":0,"down_at":14,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6801","nonce":2178563013}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6803","nonce":2178563013}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6807","nonce":2178563013}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":2178563013},{"type":"v1","addr":"192.168.123.109:6805","nonce":2178563013}]},"public_addr":"192.168.123.109:6801/2178563013","cluster_addr":"192.168.123.109:6803/2178563013","heartbeat_back_addr":"192.168.123.109:6807/2178563013","heartbeat_front_addr":"192.168.123.109:6805/2178563013","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-03-31T11:34:05.374891+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-03-31T11:34:12.966607+0000","dead_epoch":14}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.106:0/1709057529":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/1073351636":"2026-04-01T11:33:34.349039+0000","192.168.123.106:6801/947852417":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/1554541810":"2026-04-01T11:33:22.475137+0000","192.168.123.106:0/708176341":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/2978357861":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6801/942312008":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6800/947852417":"2026-04-01T11:33:34.349039+0000","192.168.123.106:0/3320458621":"2026-04-01T11:33:22.475137+0000","192.168.123.106:6800/942312008":"2026-04-01T11:33:22.475137+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-31T11:36:18.136 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph tell osd.0 flush_pg_stats 2026-03-31T11:36:18.136 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph tell osd.1 flush_pg_stats 2026-03-31T11:36:18.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:18 vm09 bash[20661]: audit 2026-03-31T11:36:17.512545+0000 mon.a (mon.0) 436 : audit [DBG] from='client.? 192.168.123.106:0/839866189' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:18.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:18 vm09 bash[20661]: audit 2026-03-31T11:36:17.512545+0000 mon.a (mon.0) 436 : audit [DBG] from='client.? 192.168.123.106:0/839866189' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:18.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:18 vm09 bash[20661]: audit 2026-03-31T11:36:17.886081+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:18.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:18 vm09 bash[20661]: audit 2026-03-31T11:36:17.886081+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:18.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:18 vm09 bash[20661]: audit 2026-03-31T11:36:17.889571+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:18.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:18 vm09 bash[20661]: audit 2026-03-31T11:36:17.889571+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:18.366 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:18.368 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:18.382 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:18 vm06 bash[20457]: audit 2026-03-31T11:36:17.512545+0000 mon.a (mon.0) 436 : audit [DBG] from='client.? 192.168.123.106:0/839866189' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:18.382 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:18 vm06 bash[20457]: audit 2026-03-31T11:36:17.512545+0000 mon.a (mon.0) 436 : audit [DBG] from='client.? 192.168.123.106:0/839866189' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:18.382 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:18 vm06 bash[20457]: audit 2026-03-31T11:36:17.886081+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:18.382 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:18 vm06 bash[20457]: audit 2026-03-31T11:36:17.886081+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:18.382 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:18 vm06 bash[20457]: audit 2026-03-31T11:36:17.889571+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:18.382 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:18 vm06 bash[20457]: audit 2026-03-31T11:36:17.889571+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:18.621 INFO:teuthology.orchestra.run.vm06.stdout:68719476738 2026-03-31T11:36:18.621 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd last-stat-seq osd.1 2026-03-31T11:36:18.659 INFO:teuthology.orchestra.run.vm06.stdout:34359738396 2026-03-31T11:36:18.659 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd last-stat-seq osd.0 2026-03-31T11:36:18.855 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:18.937 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:19.154 INFO:teuthology.orchestra.run.vm06.stdout:1 2026-03-31T11:36:19.164 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:19 vm06 bash[20457]: audit 2026-03-31T11:36:18.086592+0000 mon.a (mon.0) 439 : audit [DBG] from='client.? 192.168.123.106:0/696361940' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:19.165 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:19 vm06 bash[20457]: audit 2026-03-31T11:36:18.086592+0000 mon.a (mon.0) 439 : audit [DBG] from='client.? 192.168.123.106:0/696361940' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:19.165 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:19 vm06 bash[20457]: cluster 2026-03-31T11:36:18.147749+0000 mgr.a (mgr.14152) 150 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:19.165 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:19 vm06 bash[20457]: cluster 2026-03-31T11:36:18.147749+0000 mgr.a (mgr.14152) 150 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:19.202 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476738 got 1 for osd.1 2026-03-31T11:36:19.230 INFO:teuthology.orchestra.run.vm06.stdout:34359738395 2026-03-31T11:36:19.278 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738396 got 34359738395 for osd.0 2026-03-31T11:36:19.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:19 vm09 bash[20661]: audit 2026-03-31T11:36:18.086592+0000 mon.a (mon.0) 439 : audit [DBG] from='client.? 192.168.123.106:0/696361940' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:19.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:19 vm09 bash[20661]: audit 2026-03-31T11:36:18.086592+0000 mon.a (mon.0) 439 : audit [DBG] from='client.? 192.168.123.106:0/696361940' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-31T11:36:19.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:19 vm09 bash[20661]: cluster 2026-03-31T11:36:18.147749+0000 mgr.a (mgr.14152) 150 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:19.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:19 vm09 bash[20661]: cluster 2026-03-31T11:36:18.147749+0000 mgr.a (mgr.14152) 150 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:20.203 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd last-stat-seq osd.1 2026-03-31T11:36:20.279 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd last-stat-seq osd.0 2026-03-31T11:36:20.392 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:20 vm09 bash[20661]: audit 2026-03-31T11:36:19.154922+0000 mon.a (mon.0) 440 : audit [DBG] from='client.? 192.168.123.106:0/1423521092' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:20.392 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:20 vm09 bash[20661]: audit 2026-03-31T11:36:19.154922+0000 mon.a (mon.0) 440 : audit [DBG] from='client.? 192.168.123.106:0/1423521092' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:20.392 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:20 vm09 bash[20661]: audit 2026-03-31T11:36:19.230651+0000 mon.a (mon.0) 441 : audit [DBG] from='client.? 192.168.123.106:0/1953444856' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:20.392 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:20 vm09 bash[20661]: audit 2026-03-31T11:36:19.230651+0000 mon.a (mon.0) 441 : audit [DBG] from='client.? 192.168.123.106:0/1953444856' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:20.440 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:20.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:20 vm06 bash[20457]: audit 2026-03-31T11:36:19.154922+0000 mon.a (mon.0) 440 : audit [DBG] from='client.? 192.168.123.106:0/1423521092' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:20.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:20 vm06 bash[20457]: audit 2026-03-31T11:36:19.154922+0000 mon.a (mon.0) 440 : audit [DBG] from='client.? 192.168.123.106:0/1423521092' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:20.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:20 vm06 bash[20457]: audit 2026-03-31T11:36:19.230651+0000 mon.a (mon.0) 441 : audit [DBG] from='client.? 192.168.123.106:0/1953444856' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:20.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:20 vm06 bash[20457]: audit 2026-03-31T11:36:19.230651+0000 mon.a (mon.0) 441 : audit [DBG] from='client.? 192.168.123.106:0/1953444856' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:20.529 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:20.750 INFO:teuthology.orchestra.run.vm06.stdout:1 2026-03-31T11:36:20.822 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476738 got 1 for osd.1 2026-03-31T11:36:20.847 INFO:teuthology.orchestra.run.vm06.stdout:34359738395 2026-03-31T11:36:20.896 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738396 got 34359738395 for osd.0 2026-03-31T11:36:21.452 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:21 vm06 bash[20457]: cluster 2026-03-31T11:36:20.147973+0000 mgr.a (mgr.14152) 151 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:21.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:21 vm06 bash[20457]: cluster 2026-03-31T11:36:20.147973+0000 mgr.a (mgr.14152) 151 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:21.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:21 vm06 bash[20457]: audit 2026-03-31T11:36:20.750041+0000 mon.a (mon.0) 442 : audit [DBG] from='client.? 192.168.123.106:0/1376012763' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:21.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:21 vm06 bash[20457]: audit 2026-03-31T11:36:20.750041+0000 mon.a (mon.0) 442 : audit [DBG] from='client.? 192.168.123.106:0/1376012763' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:21.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:21 vm06 bash[20457]: audit 2026-03-31T11:36:20.847385+0000 mon.a (mon.0) 443 : audit [DBG] from='client.? 192.168.123.106:0/1009662815' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:21.453 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:21 vm06 bash[20457]: audit 2026-03-31T11:36:20.847385+0000 mon.a (mon.0) 443 : audit [DBG] from='client.? 192.168.123.106:0/1009662815' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:21.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:21 vm09 bash[20661]: cluster 2026-03-31T11:36:20.147973+0000 mgr.a (mgr.14152) 151 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:21.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:21 vm09 bash[20661]: cluster 2026-03-31T11:36:20.147973+0000 mgr.a (mgr.14152) 151 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:21.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:21 vm09 bash[20661]: audit 2026-03-31T11:36:20.750041+0000 mon.a (mon.0) 442 : audit [DBG] from='client.? 192.168.123.106:0/1376012763' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:21.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:21 vm09 bash[20661]: audit 2026-03-31T11:36:20.750041+0000 mon.a (mon.0) 442 : audit [DBG] from='client.? 192.168.123.106:0/1376012763' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:21.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:21 vm09 bash[20661]: audit 2026-03-31T11:36:20.847385+0000 mon.a (mon.0) 443 : audit [DBG] from='client.? 192.168.123.106:0/1009662815' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:21.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:21 vm09 bash[20661]: audit 2026-03-31T11:36:20.847385+0000 mon.a (mon.0) 443 : audit [DBG] from='client.? 192.168.123.106:0/1009662815' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:21.822 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd last-stat-seq osd.1 2026-03-31T11:36:21.897 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph osd last-stat-seq osd.0 2026-03-31T11:36:22.060 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:22.153 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:22.359 INFO:teuthology.orchestra.run.vm06.stdout:68719476739 2026-03-31T11:36:22.445 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476738 got 68719476739 for osd.1 2026-03-31T11:36:22.445 DEBUG:teuthology.parallel:result is None 2026-03-31T11:36:22.460 INFO:teuthology.orchestra.run.vm06.stdout:34359738396 2026-03-31T11:36:22.534 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738396 got 34359738396 for osd.0 2026-03-31T11:36:22.534 DEBUG:teuthology.parallel:result is None 2026-03-31T11:36:22.534 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-31T11:36:22.534 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph pg dump --format=json 2026-03-31T11:36:22.812 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:23.106 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:23.106 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.047240+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.047240+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.050325+0000 mon.a (mon.0) 445 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.050325+0000 mon.a (mon.0) 445 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: cephadm 2026-03-31T11:36:22.051115+0000 mgr.a (mgr.14152) 152 : cephadm [INF] Detected new or changed devices on vm09 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: cephadm 2026-03-31T11:36:22.051115+0000 mgr.a (mgr.14152) 152 : cephadm [INF] Detected new or changed devices on vm09 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: cluster 2026-03-31T11:36:22.051371+0000 mgr.a (mgr.14152) 153 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: cluster 2026-03-31T11:36:22.051371+0000 mgr.a (mgr.14152) 153 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.116 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.052122+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.052122+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.053389+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.053389+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.053860+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.053860+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.054731+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.054731+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: cluster 2026-03-31T11:36:22.056068+0000 mgr.a (mgr.14152) 154 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: cluster 2026-03-31T11:36:22.056068+0000 mgr.a (mgr.14152) 154 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: cluster 2026-03-31T11:36:22.056117+0000 mgr.a (mgr.14152) 155 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: cluster 2026-03-31T11:36:22.056117+0000 mgr.a (mgr.14152) 155 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.058479+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.058479+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.059431+0000 mon.a (mon.0) 451 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.059431+0000 mon.a (mon.0) 451 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.060424+0000 mon.a (mon.0) 452 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.060424+0000 mon.a (mon.0) 452 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.060768+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.060768+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.359350+0000 mon.a (mon.0) 454 : audit [DBG] from='client.? 192.168.123.106:0/1613942410' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.359350+0000 mon.a (mon.0) 454 : audit [DBG] from='client.? 192.168.123.106:0/1613942410' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.460824+0000 mon.a (mon.0) 455 : audit [DBG] from='client.? 192.168.123.106:0/2507253147' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:23.117 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:23 vm06 bash[20457]: audit 2026-03-31T11:36:22.460824+0000 mon.a (mon.0) 455 : audit [DBG] from='client.? 192.168.123.106:0/2507253147' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:23.156 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":103,"stamp":"2026-03-31T11:36:22.056099+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":54416,"kb_used_data":272,"kb_used_omap":14,"kb_used_meta":54065,"kb_avail":41880432,"statfs":{"total":42941284352,"available":42885562368,"internally_reserved":0,"allocated":278528,"data_stored":73508,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":15109,"internal_metadata":55362811},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":16,"seq":68719476739,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27432,"kb_used_data":136,"kb_used_omap":5,"kb_used_meta":27258,"kb_avail":20939992,"statfs":{"total":21470642176,"available":21442551808,"internally_reserved":0,"allocated":139264,"data_stored":36754,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":5688,"internal_metadata":27912648},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738396,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26984,"kb_used_data":136,"kb_used_omap":9,"kb_used_meta":26806,"kb_avail":20940440,"statfs":{"total":21470642176,"available":21443010560,"internally_reserved":0,"allocated":139264,"data_stored":36754,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":9421,"internal_metadata":27450163},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-31T11:36:23.156 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph pg dump --format=json 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.047240+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.047240+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.050325+0000 mon.a (mon.0) 445 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.050325+0000 mon.a (mon.0) 445 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: cephadm 2026-03-31T11:36:22.051115+0000 mgr.a (mgr.14152) 152 : cephadm [INF] Detected new or changed devices on vm09 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: cephadm 2026-03-31T11:36:22.051115+0000 mgr.a (mgr.14152) 152 : cephadm [INF] Detected new or changed devices on vm09 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: cluster 2026-03-31T11:36:22.051371+0000 mgr.a (mgr.14152) 153 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: cluster 2026-03-31T11:36:22.051371+0000 mgr.a (mgr.14152) 153 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.052122+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.052122+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.053389+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.053389+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.053860+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.053860+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:23.224 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.054731+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.054731+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: cluster 2026-03-31T11:36:22.056068+0000 mgr.a (mgr.14152) 154 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: cluster 2026-03-31T11:36:22.056068+0000 mgr.a (mgr.14152) 154 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: cluster 2026-03-31T11:36:22.056117+0000 mgr.a (mgr.14152) 155 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: cluster 2026-03-31T11:36:22.056117+0000 mgr.a (mgr.14152) 155 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.058479+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.058479+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.059431+0000 mon.a (mon.0) 451 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.059431+0000 mon.a (mon.0) 451 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.060424+0000 mon.a (mon.0) 452 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.060424+0000 mon.a (mon.0) 452 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.060768+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.060768+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.359350+0000 mon.a (mon.0) 454 : audit [DBG] from='client.? 192.168.123.106:0/1613942410' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.359350+0000 mon.a (mon.0) 454 : audit [DBG] from='client.? 192.168.123.106:0/1613942410' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.460824+0000 mon.a (mon.0) 455 : audit [DBG] from='client.? 192.168.123.106:0/2507253147' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:23.225 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:23 vm09 bash[20661]: audit 2026-03-31T11:36:22.460824+0000 mon.a (mon.0) 455 : audit [DBG] from='client.? 192.168.123.106:0/2507253147' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-31T11:36:23.398 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:23.677 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-03-31T11:36:23.677 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:23.725 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":103,"stamp":"2026-03-31T11:36:22.056099+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":54416,"kb_used_data":272,"kb_used_omap":14,"kb_used_meta":54065,"kb_avail":41880432,"statfs":{"total":42941284352,"available":42885562368,"internally_reserved":0,"allocated":278528,"data_stored":73508,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":15109,"internal_metadata":55362811},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":16,"seq":68719476739,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27432,"kb_used_data":136,"kb_used_omap":5,"kb_used_meta":27258,"kb_avail":20939992,"statfs":{"total":21470642176,"available":21442551808,"internally_reserved":0,"allocated":139264,"data_stored":36754,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":5688,"internal_metadata":27912648},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738396,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26984,"kb_used_data":136,"kb_used_omap":9,"kb_used_meta":26806,"kb_avail":20940440,"statfs":{"total":21470642176,"available":21443010560,"internally_reserved":0,"allocated":139264,"data_stored":36754,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":9421,"internal_metadata":27450163},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-31T11:36:23.725 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-31T11:36:23.725 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-31T11:36:23.725 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-31T11:36:23.725 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph health --format=json 2026-03-31T11:36:23.960 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:24.281 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:24.281 INFO:teuthology.orchestra.run.vm06.stdout:{"status":"HEALTH_WARN","checks":{"CEPHADM_FAILED_DAEMON":{"severity":"HEALTH_WARN","summary":{"message":"1 failed cephadm daemon(s)","count":1},"muted":false}},"mutes":[]} 2026-03-31T11:36:24.291 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:24 vm06 bash[20457]: cluster 2026-03-31T11:36:23.051207+0000 mon.a (mon.0) 456 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-31T11:36:24.291 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:24 vm06 bash[20457]: cluster 2026-03-31T11:36:23.051207+0000 mon.a (mon.0) 456 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-31T11:36:24.291 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:24 vm06 bash[20457]: audit 2026-03-31T11:36:23.106197+0000 mgr.a (mgr.14152) 156 : audit [DBG] from='client.14308 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:24.291 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:24 vm06 bash[20457]: audit 2026-03-31T11:36:23.106197+0000 mgr.a (mgr.14152) 156 : audit [DBG] from='client.14308 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:24.291 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:24 vm06 bash[20457]: audit 2026-03-31T11:36:23.676994+0000 mgr.a (mgr.14152) 157 : audit [DBG] from='client.14312 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:24.291 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:24 vm06 bash[20457]: audit 2026-03-31T11:36:23.676994+0000 mgr.a (mgr.14152) 157 : audit [DBG] from='client.14312 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:24.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:24 vm09 bash[20661]: cluster 2026-03-31T11:36:23.051207+0000 mon.a (mon.0) 456 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-31T11:36:24.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:24 vm09 bash[20661]: cluster 2026-03-31T11:36:23.051207+0000 mon.a (mon.0) 456 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-31T11:36:24.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:24 vm09 bash[20661]: audit 2026-03-31T11:36:23.106197+0000 mgr.a (mgr.14152) 156 : audit [DBG] from='client.14308 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:24.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:24 vm09 bash[20661]: audit 2026-03-31T11:36:23.106197+0000 mgr.a (mgr.14152) 156 : audit [DBG] from='client.14308 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:24.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:24 vm09 bash[20661]: audit 2026-03-31T11:36:23.676994+0000 mgr.a (mgr.14152) 157 : audit [DBG] from='client.14312 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:24.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:24 vm09 bash[20661]: audit 2026-03-31T11:36:23.676994+0000 mgr.a (mgr.14152) 157 : audit [DBG] from='client.14312 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:25.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:25 vm06 bash[20457]: cluster 2026-03-31T11:36:24.056306+0000 mgr.a (mgr.14152) 158 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:25.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:25 vm06 bash[20457]: cluster 2026-03-31T11:36:24.056306+0000 mgr.a (mgr.14152) 158 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:25.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:25 vm06 bash[20457]: audit 2026-03-31T11:36:24.281016+0000 mon.a (mon.0) 457 : audit [DBG] from='client.? 192.168.123.106:0/1936290522' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-31T11:36:25.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:25 vm06 bash[20457]: audit 2026-03-31T11:36:24.281016+0000 mon.a (mon.0) 457 : audit [DBG] from='client.? 192.168.123.106:0/1936290522' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-31T11:36:25.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:25 vm06 bash[20457]: audit 2026-03-31T11:36:24.892656+0000 mon.a (mon.0) 458 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:25.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:25 vm06 bash[20457]: audit 2026-03-31T11:36:24.892656+0000 mon.a (mon.0) 458 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:25.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:25 vm06 bash[20457]: audit 2026-03-31T11:36:24.895742+0000 mon.a (mon.0) 459 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:25.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:25 vm06 bash[20457]: audit 2026-03-31T11:36:24.895742+0000 mon.a (mon.0) 459 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:25 vm09 bash[20661]: cluster 2026-03-31T11:36:24.056306+0000 mgr.a (mgr.14152) 158 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:25 vm09 bash[20661]: cluster 2026-03-31T11:36:24.056306+0000 mgr.a (mgr.14152) 158 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:25 vm09 bash[20661]: audit 2026-03-31T11:36:24.281016+0000 mon.a (mon.0) 457 : audit [DBG] from='client.? 192.168.123.106:0/1936290522' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-31T11:36:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:25 vm09 bash[20661]: audit 2026-03-31T11:36:24.281016+0000 mon.a (mon.0) 457 : audit [DBG] from='client.? 192.168.123.106:0/1936290522' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-31T11:36:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:25 vm09 bash[20661]: audit 2026-03-31T11:36:24.892656+0000 mon.a (mon.0) 458 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:25 vm09 bash[20661]: audit 2026-03-31T11:36:24.892656+0000 mon.a (mon.0) 458 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:25 vm09 bash[20661]: audit 2026-03-31T11:36:24.895742+0000 mon.a (mon.0) 459 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:25.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:25 vm09 bash[20661]: audit 2026-03-31T11:36:24.895742+0000 mon.a (mon.0) 459 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.490778+0000 mon.a (mon.0) 460 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.490778+0000 mon.a (mon.0) 460 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.493533+0000 mon.a (mon.0) 461 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.493533+0000 mon.a (mon.0) 461 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: cluster 2026-03-31T11:36:25.493866+0000 mgr.a (mgr.14152) 159 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: cluster 2026-03-31T11:36:25.493866+0000 mgr.a (mgr.14152) 159 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.494962+0000 mon.a (mon.0) 462 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.494962+0000 mon.a (mon.0) 462 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.496021+0000 mon.a (mon.0) 463 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.496021+0000 mon.a (mon.0) 463 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.497146+0000 mon.a (mon.0) 464 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.497146+0000 mon.a (mon.0) 464 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.497384+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.497384+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.500321+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.500321+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.501313+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:26.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:26 vm06 bash[20457]: audit 2026-03-31T11:36:25.501313+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:26.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.490778+0000 mon.a (mon.0) 460 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.490778+0000 mon.a (mon.0) 460 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.493533+0000 mon.a (mon.0) 461 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.493533+0000 mon.a (mon.0) 461 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: cluster 2026-03-31T11:36:25.493866+0000 mgr.a (mgr.14152) 159 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: cluster 2026-03-31T11:36:25.493866+0000 mgr.a (mgr.14152) 159 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.494962+0000 mon.a (mon.0) 462 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.494962+0000 mon.a (mon.0) 462 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.496021+0000 mon.a (mon.0) 463 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.496021+0000 mon.a (mon.0) 463 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.497146+0000 mon.a (mon.0) 464 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.497146+0000 mon.a (mon.0) 464 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.497384+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.497384+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.500321+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.500321+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.501313+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:26.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:26 vm09 bash[20661]: audit 2026-03-31T11:36:25.501313+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:27.331 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph health --format=json 2026-03-31T11:36:27.561 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: cluster 2026-03-31T11:36:26.495643+0000 mon.a (mon.0) 468 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: cluster 2026-03-31T11:36:26.495643+0000 mon.a (mon.0) 468 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: cluster 2026-03-31T11:36:26.495655+0000 mon.a (mon.0) 469 : cluster [INF] Cluster is now healthy 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: cluster 2026-03-31T11:36:26.495655+0000 mon.a (mon.0) 469 : cluster [INF] Cluster is now healthy 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: audit 2026-03-31T11:36:26.939500+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: audit 2026-03-31T11:36:26.939500+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: audit 2026-03-31T11:36:26.942394+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: audit 2026-03-31T11:36:26.942394+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: audit 2026-03-31T11:36:26.945623+0000 mon.a (mon.0) 472 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.600 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:27 vm06 bash[20457]: audit 2026-03-31T11:36:26.945623+0000 mon.a (mon.0) 472 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: cluster 2026-03-31T11:36:26.495643+0000 mon.a (mon.0) 468 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-31T11:36:27.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: cluster 2026-03-31T11:36:26.495643+0000 mon.a (mon.0) 468 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-31T11:36:27.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: cluster 2026-03-31T11:36:26.495655+0000 mon.a (mon.0) 469 : cluster [INF] Cluster is now healthy 2026-03-31T11:36:27.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: cluster 2026-03-31T11:36:26.495655+0000 mon.a (mon.0) 469 : cluster [INF] Cluster is now healthy 2026-03-31T11:36:27.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: audit 2026-03-31T11:36:26.939500+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: audit 2026-03-31T11:36:26.939500+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: audit 2026-03-31T11:36:26.942394+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: audit 2026-03-31T11:36:26.942394+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: audit 2026-03-31T11:36:26.945623+0000 mon.a (mon.0) 472 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:27 vm09 bash[20661]: audit 2026-03-31T11:36:26.945623+0000 mon.a (mon.0) 472 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:27.865 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:27.865 INFO:teuthology.orchestra.run.vm06.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-31T11:36:27.910 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-31T11:36:27.910 INFO:tasks.cephadm:Setup complete, yielding 2026-03-31T11:36:27.910 INFO:teuthology.run_tasks:Running task exec... 2026-03-31T11:36:27.912 INFO:teuthology.task.exec:Executing custom commands... 2026-03-31T11:36:27.912 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm06.local 2026-03-31T11:36:27.913 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-31T11:36:27.920 INFO:teuthology.task.exec:Running commands on role host.b host ubuntu@vm09.local 2026-03-31T11:36:27.920 DEBUG:teuthology.orchestra.run.vm09:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-31T11:36:27.929 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-31T11:36:27.932 INFO:tasks.cephadm:Applying spec(s): extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon --- custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest --- custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z 2026-03-31T11:36:27.932 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch apply -i - 2026-03-31T11:36:28.197 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:28.506 INFO:teuthology.orchestra.run.vm06.stdout:Scheduled mon update... 2026-03-31T11:36:28.506 INFO:teuthology.orchestra.run.vm06.stdout:Scheduled container.foo update... 2026-03-31T11:36:28.506 INFO:teuthology.orchestra.run.vm06.stdout:Scheduled container.bar update... 2026-03-31T11:36:28.560 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-31T11:36:28.562 INFO:tasks.cephadm:Waiting for ceph service mon to start (timeout 300)... 2026-03-31T11:36:28.562 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:28.752 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:28 vm06 bash[20457]: cluster 2026-03-31T11:36:27.494062+0000 mgr.a (mgr.14152) 160 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:28.752 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:28 vm06 bash[20457]: cluster 2026-03-31T11:36:27.494062+0000 mgr.a (mgr.14152) 160 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:28.752 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:28 vm06 bash[20457]: audit 2026-03-31T11:36:27.865019+0000 mon.a (mon.0) 473 : audit [DBG] from='client.? 192.168.123.106:0/1703647588' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-31T11:36:28.753 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:28 vm06 bash[20457]: audit 2026-03-31T11:36:27.865019+0000 mon.a (mon.0) 473 : audit [DBG] from='client.? 192.168.123.106:0/1703647588' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-31T11:36:28.753 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:28 vm06 bash[20457]: audit 2026-03-31T11:36:28.494135+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:28.753 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:28 vm06 bash[20457]: audit 2026-03-31T11:36:28.494135+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:28.795 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:28.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:28 vm09 bash[20661]: cluster 2026-03-31T11:36:27.494062+0000 mgr.a (mgr.14152) 160 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:28.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:28 vm09 bash[20661]: cluster 2026-03-31T11:36:27.494062+0000 mgr.a (mgr.14152) 160 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:28.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:28 vm09 bash[20661]: audit 2026-03-31T11:36:27.865019+0000 mon.a (mon.0) 473 : audit [DBG] from='client.? 192.168.123.106:0/1703647588' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-31T11:36:28.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:28 vm09 bash[20661]: audit 2026-03-31T11:36:27.865019+0000 mon.a (mon.0) 473 : audit [DBG] from='client.? 192.168.123.106:0/1703647588' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-31T11:36:28.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:28 vm09 bash[20661]: audit 2026-03-31T11:36:28.494135+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:28.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:28 vm09 bash[20661]: audit 2026-03-31T11:36:28.494135+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:28 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.065 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:36:28 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.066 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:36:28 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.348 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:29.348 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:25.486766Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.181880Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:25.486704Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:25.486746Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:25.486796Z", "running": 2, "size": 2}}] 2026-03-31T11:36:29.359 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.359 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:36:29 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.359 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:36:29 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.398 INFO:tasks.cephadm:mon has 2/2 2026-03-31T11:36:29.398 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-31T11:36:29.400 INFO:tasks.cephadm:Waiting for ceph service container.foo to start (timeout 300)... 2026-03-31T11:36:29.400 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:29.644 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.485054+0000 mgr.a (mgr.14152) 161 : audit [DBG] from='client.14324 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.485054+0000 mgr.a (mgr.14152) 161 : audit [DBG] from='client.14324 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: cephadm 2026-03-31T11:36:28.490141+0000 mgr.a (mgr.14152) 162 : cephadm [INF] Saving service mon spec with placement * 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: cephadm 2026-03-31T11:36:28.490141+0000 mgr.a (mgr.14152) 162 : cephadm [INF] Saving service mon spec with placement * 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: cephadm 2026-03-31T11:36:28.494545+0000 mgr.a (mgr.14152) 163 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: cephadm 2026-03-31T11:36:28.494545+0000 mgr.a (mgr.14152) 163 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.494762+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.494762+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.496349+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.496349+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.500348+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.500348+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.501985+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.501985+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: cephadm 2026-03-31T11:36:28.502344+0000 mgr.a (mgr.14152) 164 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: cephadm 2026-03-31T11:36:28.502344+0000 mgr.a (mgr.14152) 164 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.503940+0000 mon.a (mon.0) 479 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.503940+0000 mon.a (mon.0) 479 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.506459+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.506459+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.508489+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.508489+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.509236+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:28.509236+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: cephadm 2026-03-31T11:36:28.510890+0000 mgr.a (mgr.14152) 165 : cephadm [INF] Deploying daemon container.foo.vm06 on vm06 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: cephadm 2026-03-31T11:36:28.510890+0000 mgr.a (mgr.14152) 165 : cephadm [INF] Deploying daemon container.foo.vm06 on vm06 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:29.175441+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:29.175441+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:29.178649+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:29.178649+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:29.181743+0000 mon.a (mon.0) 485 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.659 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:29 vm06 bash[20457]: audit 2026-03-31T11:36:29.181743+0000 mon.a (mon.0) 485 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:36:29 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.766 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:36:29 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.485054+0000 mgr.a (mgr.14152) 161 : audit [DBG] from='client.14324 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.485054+0000 mgr.a (mgr.14152) 161 : audit [DBG] from='client.14324 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: cephadm 2026-03-31T11:36:28.490141+0000 mgr.a (mgr.14152) 162 : cephadm [INF] Saving service mon spec with placement * 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: cephadm 2026-03-31T11:36:28.490141+0000 mgr.a (mgr.14152) 162 : cephadm [INF] Saving service mon spec with placement * 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: cephadm 2026-03-31T11:36:28.494545+0000 mgr.a (mgr.14152) 163 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: cephadm 2026-03-31T11:36:28.494545+0000 mgr.a (mgr.14152) 163 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.494762+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.494762+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.496349+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.496349+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.500348+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.500348+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.501985+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.501985+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: cephadm 2026-03-31T11:36:28.502344+0000 mgr.a (mgr.14152) 164 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: cephadm 2026-03-31T11:36:28.502344+0000 mgr.a (mgr.14152) 164 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.503940+0000 mon.a (mon.0) 479 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.503940+0000 mon.a (mon.0) 479 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.506459+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.766 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.506459+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.508489+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.508489+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.509236+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:28.509236+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: cephadm 2026-03-31T11:36:28.510890+0000 mgr.a (mgr.14152) 165 : cephadm [INF] Deploying daemon container.foo.vm06 on vm06 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: cephadm 2026-03-31T11:36:28.510890+0000 mgr.a (mgr.14152) 165 : cephadm [INF] Deploying daemon container.foo.vm06 on vm06 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:29.175441+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:29.175441+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:29.178649+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:29.178649+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:29.181743+0000 mon.a (mon.0) 485 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 bash[20661]: audit 2026-03-31T11:36:29.181743+0000 mon.a (mon.0) 485 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:29.767 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:29.947 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:29.947 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:25.486766Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:25.486704Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:25.486746Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:25.486796Z", "running": 2, "size": 2}}] 2026-03-31T11:36:30.015 INFO:tasks.cephadm:container.foo has 0/2 2026-03-31T11:36:30.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:29 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.085 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:36:29 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.085 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:36:29 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.460 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.460 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:36:30 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.460 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:36:30 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: cephadm 2026-03-31T11:36:29.182320+0000 mgr.a (mgr.14152) 166 : cephadm [INF] Deploying daemon container.foo.vm09 on vm09 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: cephadm 2026-03-31T11:36:29.182320+0000 mgr.a (mgr.14152) 166 : cephadm [INF] Deploying daemon container.foo.vm09 on vm09 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.347392+0000 mgr.a (mgr.14152) 167 : audit [DBG] from='client.14328 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.347392+0000 mgr.a (mgr.14152) 167 : audit [DBG] from='client.14328 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: cluster 2026-03-31T11:36:29.494283+0000 mgr.a (mgr.14152) 168 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: cluster 2026-03-31T11:36:29.494283+0000 mgr.a (mgr.14152) 168 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.867780+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.867780+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.870626+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.870626+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.873436+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.873436+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.876089+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.658 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:30 vm06 bash[20457]: audit 2026-03-31T11:36:29.876089+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.834 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:36:30 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.834 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:36:30 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: cephadm 2026-03-31T11:36:29.182320+0000 mgr.a (mgr.14152) 166 : cephadm [INF] Deploying daemon container.foo.vm09 on vm09 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: cephadm 2026-03-31T11:36:29.182320+0000 mgr.a (mgr.14152) 166 : cephadm [INF] Deploying daemon container.foo.vm09 on vm09 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.347392+0000 mgr.a (mgr.14152) 167 : audit [DBG] from='client.14328 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.347392+0000 mgr.a (mgr.14152) 167 : audit [DBG] from='client.14328 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: cluster 2026-03-31T11:36:29.494283+0000 mgr.a (mgr.14152) 168 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: cluster 2026-03-31T11:36:29.494283+0000 mgr.a (mgr.14152) 168 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.867780+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.867780+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.870626+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.870626+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.873436+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.873436+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.876089+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:30.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:30 vm09 bash[20661]: audit 2026-03-31T11:36:29.876089+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:31.015 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:31.267 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:31.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:31 vm06 bash[20457]: cephadm 2026-03-31T11:36:29.877432+0000 mgr.a (mgr.14152) 169 : cephadm [INF] Deploying daemon container.bar.vm09 on vm09 2026-03-31T11:36:31.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:31 vm06 bash[20457]: cephadm 2026-03-31T11:36:29.877432+0000 mgr.a (mgr.14152) 169 : cephadm [INF] Deploying daemon container.bar.vm09 on vm09 2026-03-31T11:36:31.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:31 vm06 bash[20457]: audit 2026-03-31T11:36:29.945419+0000 mgr.a (mgr.14152) 170 : audit [DBG] from='client.14332 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:31.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:31 vm06 bash[20457]: audit 2026-03-31T11:36:29.945419+0000 mgr.a (mgr.14152) 170 : audit [DBG] from='client.14332 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:31.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:31 vm06 bash[20457]: audit 2026-03-31T11:36:31.060389+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:31.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:31 vm06 bash[20457]: audit 2026-03-31T11:36:31.060389+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:31.570 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:31.570 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:25.486766Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:25.486704Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:25.486746Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:25.486796Z", "running": 2, "size": 2}}] 2026-03-31T11:36:31.635 INFO:tasks.cephadm:container.foo has 0/2 2026-03-31T11:36:31.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:31 vm09 bash[20661]: cephadm 2026-03-31T11:36:29.877432+0000 mgr.a (mgr.14152) 169 : cephadm [INF] Deploying daemon container.bar.vm09 on vm09 2026-03-31T11:36:31.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:31 vm09 bash[20661]: cephadm 2026-03-31T11:36:29.877432+0000 mgr.a (mgr.14152) 169 : cephadm [INF] Deploying daemon container.bar.vm09 on vm09 2026-03-31T11:36:31.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:31 vm09 bash[20661]: audit 2026-03-31T11:36:29.945419+0000 mgr.a (mgr.14152) 170 : audit [DBG] from='client.14332 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:31.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:31 vm09 bash[20661]: audit 2026-03-31T11:36:29.945419+0000 mgr.a (mgr.14152) 170 : audit [DBG] from='client.14332 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:31.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:31 vm09 bash[20661]: audit 2026-03-31T11:36:31.060389+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:31.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:31 vm09 bash[20661]: audit 2026-03-31T11:36:31.060389+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:32.636 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:32.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:32 vm06 bash[20457]: cluster 2026-03-31T11:36:31.494520+0000 mgr.a (mgr.14152) 171 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:32.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:32 vm06 bash[20457]: cluster 2026-03-31T11:36:31.494520+0000 mgr.a (mgr.14152) 171 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:32.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:32 vm06 bash[20457]: audit 2026-03-31T11:36:31.569933+0000 mgr.a (mgr.14152) 172 : audit [DBG] from='client.14336 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:32.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:32 vm06 bash[20457]: audit 2026-03-31T11:36:31.569933+0000 mgr.a (mgr.14152) 172 : audit [DBG] from='client.14336 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:32.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:32 vm06 bash[20457]: audit 2026-03-31T11:36:32.355477+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:32.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:32 vm06 bash[20457]: audit 2026-03-31T11:36:32.355477+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:32.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:32 vm09 bash[20661]: cluster 2026-03-31T11:36:31.494520+0000 mgr.a (mgr.14152) 171 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:32.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:32 vm09 bash[20661]: cluster 2026-03-31T11:36:31.494520+0000 mgr.a (mgr.14152) 171 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:32.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:32 vm09 bash[20661]: audit 2026-03-31T11:36:31.569933+0000 mgr.a (mgr.14152) 172 : audit [DBG] from='client.14336 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:32.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:32 vm09 bash[20661]: audit 2026-03-31T11:36:31.569933+0000 mgr.a (mgr.14152) 172 : audit [DBG] from='client.14336 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:32.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:32 vm09 bash[20661]: audit 2026-03-31T11:36:32.355477+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:32.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:32 vm09 bash[20661]: audit 2026-03-31T11:36:32.355477+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:32.872 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:33.225 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:33.225 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:25.486766Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:25.486704Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:25.486746Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:25.486796Z", "running": 2, "size": 2}}] 2026-03-31T11:36:33.278 INFO:tasks.cephadm:container.foo has 0/2 2026-03-31T11:36:34.279 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:34.533 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:34.545 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:34 vm06 bash[20457]: audit 2026-03-31T11:36:33.224200+0000 mgr.a (mgr.14152) 173 : audit [DBG] from='client.14340 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:34.545 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:34 vm06 bash[20457]: audit 2026-03-31T11:36:33.224200+0000 mgr.a (mgr.14152) 173 : audit [DBG] from='client.14340 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:34.545 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:34 vm06 bash[20457]: cluster 2026-03-31T11:36:33.494777+0000 mgr.a (mgr.14152) 174 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:34.545 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:34 vm06 bash[20457]: cluster 2026-03-31T11:36:33.494777+0000 mgr.a (mgr.14152) 174 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:34.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:34 vm09 bash[20661]: audit 2026-03-31T11:36:33.224200+0000 mgr.a (mgr.14152) 173 : audit [DBG] from='client.14340 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:34.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:34 vm09 bash[20661]: audit 2026-03-31T11:36:33.224200+0000 mgr.a (mgr.14152) 173 : audit [DBG] from='client.14340 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:34.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:34 vm09 bash[20661]: cluster 2026-03-31T11:36:33.494777+0000 mgr.a (mgr.14152) 174 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:34.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:34 vm09 bash[20661]: cluster 2026-03-31T11:36:33.494777+0000 mgr.a (mgr.14152) 174 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:34.850 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:34.850 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:25.486766Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:25.486704Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:25.486746Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:25.486796Z", "running": 2, "size": 2}}] 2026-03-31T11:36:34.934 INFO:tasks.cephadm:container.foo has 0/2 2026-03-31T11:36:35.934 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:36.176 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:36.459 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:36.459 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:25.486766Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:25.486704Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:25.486746Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:25.486796Z", "running": 2, "size": 2}}] 2026-03-31T11:36:36.508 INFO:tasks.cephadm:container.foo has 0/2 2026-03-31T11:36:36.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:36 vm06 bash[20457]: audit 2026-03-31T11:36:34.849236+0000 mgr.a (mgr.14152) 175 : audit [DBG] from='client.14344 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:36.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:36 vm06 bash[20457]: audit 2026-03-31T11:36:34.849236+0000 mgr.a (mgr.14152) 175 : audit [DBG] from='client.14344 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:36.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:36 vm06 bash[20457]: cluster 2026-03-31T11:36:35.495013+0000 mgr.a (mgr.14152) 176 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:36.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:36 vm06 bash[20457]: cluster 2026-03-31T11:36:35.495013+0000 mgr.a (mgr.14152) 176 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:36.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:36 vm09 bash[20661]: audit 2026-03-31T11:36:34.849236+0000 mgr.a (mgr.14152) 175 : audit [DBG] from='client.14344 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:36.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:36 vm09 bash[20661]: audit 2026-03-31T11:36:34.849236+0000 mgr.a (mgr.14152) 175 : audit [DBG] from='client.14344 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:36.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:36 vm09 bash[20661]: cluster 2026-03-31T11:36:35.495013+0000 mgr.a (mgr.14152) 176 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:36.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:36 vm09 bash[20661]: cluster 2026-03-31T11:36:35.495013+0000 mgr.a (mgr.14152) 176 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:37.509 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:37.748 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:36.458170+0000 mgr.a (mgr.14152) 177 : audit [DBG] from='client.14348 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:36.458170+0000 mgr.a (mgr.14152) 177 : audit [DBG] from='client.14348 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.162577+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.162577+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.166187+0000 mon.a (mon.0) 493 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.166187+0000 mon.a (mon.0) 493 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.169881+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.169881+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.195870+0000 mon.a (mon.0) 495 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.195870+0000 mon.a (mon.0) 495 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.198800+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.198800+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.201906+0000 mon.a (mon.0) 497 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.790 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:37 vm06 bash[20457]: audit 2026-03-31T11:36:37.201906+0000 mon.a (mon.0) 497 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:36.458170+0000 mgr.a (mgr.14152) 177 : audit [DBG] from='client.14348 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:36.458170+0000 mgr.a (mgr.14152) 177 : audit [DBG] from='client.14348 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.162577+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.162577+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.166187+0000 mon.a (mon.0) 493 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.166187+0000 mon.a (mon.0) 493 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.169881+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.169881+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.195870+0000 mon.a (mon.0) 495 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.195870+0000 mon.a (mon.0) 495 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.198800+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.198800+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.201906+0000 mon.a (mon.0) 497 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:37.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:37 vm09 bash[20661]: audit 2026-03-31T11:36:37.201906+0000 mon.a (mon.0) 497 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:38.056 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:38.056 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:37.191674Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:38.134 INFO:tasks.cephadm:container.foo has 2/2 2026-03-31T11:36:38.134 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-31T11:36:38.136 INFO:tasks.cephadm:Waiting for ceph service container.bar to start (timeout 300)... 2026-03-31T11:36:38.136 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:38.372 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:38.666 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:38.666 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:37.191674Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:38.675 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:38 vm06 bash[20457]: cluster 2026-03-31T11:36:37.495274+0000 mgr.a (mgr.14152) 178 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:38.675 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:38 vm06 bash[20457]: cluster 2026-03-31T11:36:37.495274+0000 mgr.a (mgr.14152) 178 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:38.713 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:38.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:38 vm09 bash[20661]: cluster 2026-03-31T11:36:37.495274+0000 mgr.a (mgr.14152) 178 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:38.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:38 vm09 bash[20661]: cluster 2026-03-31T11:36:37.495274+0000 mgr.a (mgr.14152) 178 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:39.714 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:39.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:39 vm06 bash[20457]: audit 2026-03-31T11:36:38.055144+0000 mgr.a (mgr.14152) 179 : audit [DBG] from='client.14352 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:39.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:39 vm06 bash[20457]: audit 2026-03-31T11:36:38.055144+0000 mgr.a (mgr.14152) 179 : audit [DBG] from='client.14352 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:39.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:39 vm06 bash[20457]: audit 2026-03-31T11:36:38.664877+0000 mgr.a (mgr.14152) 180 : audit [DBG] from='client.14356 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:39.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:39 vm06 bash[20457]: audit 2026-03-31T11:36:38.664877+0000 mgr.a (mgr.14152) 180 : audit [DBG] from='client.14356 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:39.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:39 vm09 bash[20661]: audit 2026-03-31T11:36:38.055144+0000 mgr.a (mgr.14152) 179 : audit [DBG] from='client.14352 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:39.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:39 vm09 bash[20661]: audit 2026-03-31T11:36:38.055144+0000 mgr.a (mgr.14152) 179 : audit [DBG] from='client.14352 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:39.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:39 vm09 bash[20661]: audit 2026-03-31T11:36:38.664877+0000 mgr.a (mgr.14152) 180 : audit [DBG] from='client.14356 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:39.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:39 vm09 bash[20661]: audit 2026-03-31T11:36:38.664877+0000 mgr.a (mgr.14152) 180 : audit [DBG] from='client.14356 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:39.968 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:40.268 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:40.268 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:37.191674Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:40.325 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:40.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:40 vm06 bash[20457]: cluster 2026-03-31T11:36:39.495496+0000 mgr.a (mgr.14152) 181 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:40.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:40 vm06 bash[20457]: cluster 2026-03-31T11:36:39.495496+0000 mgr.a (mgr.14152) 181 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:40.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:40 vm09 bash[20661]: cluster 2026-03-31T11:36:39.495496+0000 mgr.a (mgr.14152) 181 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:40.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:40 vm09 bash[20661]: cluster 2026-03-31T11:36:39.495496+0000 mgr.a (mgr.14152) 181 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:41.326 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:41.574 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:41.587 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:41 vm06 bash[20457]: audit 2026-03-31T11:36:40.267102+0000 mgr.a (mgr.14152) 182 : audit [DBG] from='client.14360 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:41.587 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:41 vm06 bash[20457]: audit 2026-03-31T11:36:40.267102+0000 mgr.a (mgr.14152) 182 : audit [DBG] from='client.14360 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:41.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:41 vm09 bash[20661]: audit 2026-03-31T11:36:40.267102+0000 mgr.a (mgr.14152) 182 : audit [DBG] from='client.14360 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:41.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:41 vm09 bash[20661]: audit 2026-03-31T11:36:40.267102+0000 mgr.a (mgr.14152) 182 : audit [DBG] from='client.14360 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:42.132 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:42.132 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:37.191674Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:42.185 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:42 vm06 bash[20457]: cluster 2026-03-31T11:36:41.495755+0000 mgr.a (mgr.14152) 183 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:42 vm06 bash[20457]: cluster 2026-03-31T11:36:41.495755+0000 mgr.a (mgr.14152) 183 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:42.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:42 vm09 bash[20661]: cluster 2026-03-31T11:36:41.495755+0000 mgr.a (mgr.14152) 183 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:42.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:42 vm09 bash[20661]: cluster 2026-03-31T11:36:41.495755+0000 mgr.a (mgr.14152) 183 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:43.186 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:43.422 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:43.713 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:43.713 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:37.191674Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:43.723 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:43 vm06 bash[20457]: audit 2026-03-31T11:36:42.130842+0000 mgr.a (mgr.14152) 184 : audit [DBG] from='client.14364 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:43.723 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:43 vm06 bash[20457]: audit 2026-03-31T11:36:42.130842+0000 mgr.a (mgr.14152) 184 : audit [DBG] from='client.14364 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:43.762 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:43.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:43 vm09 bash[20661]: audit 2026-03-31T11:36:42.130842+0000 mgr.a (mgr.14152) 184 : audit [DBG] from='client.14364 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:43.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:43 vm09 bash[20661]: audit 2026-03-31T11:36:42.130842+0000 mgr.a (mgr.14152) 184 : audit [DBG] from='client.14364 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:44.763 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:44.770 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:44 vm06 bash[20457]: cluster 2026-03-31T11:36:43.496003+0000 mgr.a (mgr.14152) 185 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:44.771 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:44 vm06 bash[20457]: cluster 2026-03-31T11:36:43.496003+0000 mgr.a (mgr.14152) 185 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:44.771 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:44 vm06 bash[20457]: audit 2026-03-31T11:36:43.711694+0000 mgr.a (mgr.14152) 186 : audit [DBG] from='client.14368 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:44.771 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:44 vm06 bash[20457]: audit 2026-03-31T11:36:43.711694+0000 mgr.a (mgr.14152) 186 : audit [DBG] from='client.14368 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:44.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:44 vm09 bash[20661]: cluster 2026-03-31T11:36:43.496003+0000 mgr.a (mgr.14152) 185 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:44.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:44 vm09 bash[20661]: cluster 2026-03-31T11:36:43.496003+0000 mgr.a (mgr.14152) 185 : cluster [DBG] pgmap v114: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:44.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:44 vm09 bash[20661]: audit 2026-03-31T11:36:43.711694+0000 mgr.a (mgr.14152) 186 : audit [DBG] from='client.14368 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:44.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:44 vm09 bash[20661]: audit 2026-03-31T11:36:43.711694+0000 mgr.a (mgr.14152) 186 : audit [DBG] from='client.14368 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:45.020 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:45.345 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:45.345 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:28.506587Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:37.191674Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:45.403 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:46.404 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:46.660 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:46.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:46 vm06 bash[20457]: audit 2026-03-31T11:36:45.344080+0000 mgr.a (mgr.14152) 187 : audit [DBG] from='client.14372 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:46.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:46 vm06 bash[20457]: audit 2026-03-31T11:36:45.344080+0000 mgr.a (mgr.14152) 187 : audit [DBG] from='client.14372 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:46.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:46 vm06 bash[20457]: cluster 2026-03-31T11:36:45.496279+0000 mgr.a (mgr.14152) 188 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:46.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:46 vm06 bash[20457]: cluster 2026-03-31T11:36:45.496279+0000 mgr.a (mgr.14152) 188 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:46.775 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:46 vm09 bash[20661]: audit 2026-03-31T11:36:45.344080+0000 mgr.a (mgr.14152) 187 : audit [DBG] from='client.14372 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:46.775 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:46 vm09 bash[20661]: audit 2026-03-31T11:36:45.344080+0000 mgr.a (mgr.14152) 187 : audit [DBG] from='client.14372 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:46.775 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:46 vm09 bash[20661]: cluster 2026-03-31T11:36:45.496279+0000 mgr.a (mgr.14152) 188 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:46.775 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:46 vm09 bash[20661]: cluster 2026-03-31T11:36:45.496279+0000 mgr.a (mgr.14152) 188 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:46.973 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:46.973 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:46.740042Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:47.041 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:47.314 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:47.314 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:36:47 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:47.314 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:36:47 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:47.566 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:36:47 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:47.566 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:36:47 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:47.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:48.042 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:46.730160+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:46.730160+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:46.735788+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:46.735788+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:46.739861+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:46.739861+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: cephadm 2026-03-31T11:36:46.740497+0000 mgr.a (mgr.14152) 189 : cephadm [INF] Deploying daemon container.bar.vm06 on vm06 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: cephadm 2026-03-31T11:36:46.740497+0000 mgr.a (mgr.14152) 189 : cephadm [INF] Deploying daemon container.bar.vm06 on vm06 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:47.379661+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:47.379661+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:47.383525+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.049 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:47.383525+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.049 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:47.387550+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.049 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:47 vm06 bash[20457]: audit 2026-03-31T11:36:47.387550+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:46.730160+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:46.730160+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:46.735788+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:46.735788+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:46.739861+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:46.739861+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: cephadm 2026-03-31T11:36:46.740497+0000 mgr.a (mgr.14152) 189 : cephadm [INF] Deploying daemon container.bar.vm06 on vm06 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: cephadm 2026-03-31T11:36:46.740497+0000 mgr.a (mgr.14152) 189 : cephadm [INF] Deploying daemon container.bar.vm06 on vm06 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:47.379661+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:47.379661+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:47.383525+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:47.383525+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:47.387550+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:47 vm09 bash[20661]: audit 2026-03-31T11:36:47.387550+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:48.288 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:48.605 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:48.605 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:46.740042Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:48.658 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:49.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:48 vm06 bash[20457]: audit 2026-03-31T11:36:46.972243+0000 mgr.a (mgr.14152) 190 : audit [DBG] from='client.14376 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:49.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:48 vm06 bash[20457]: audit 2026-03-31T11:36:46.972243+0000 mgr.a (mgr.14152) 190 : audit [DBG] from='client.14376 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:49.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:48 vm06 bash[20457]: cluster 2026-03-31T11:36:47.496513+0000 mgr.a (mgr.14152) 191 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:49.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:48 vm06 bash[20457]: cluster 2026-03-31T11:36:47.496513+0000 mgr.a (mgr.14152) 191 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:49.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:48 vm09 bash[20661]: audit 2026-03-31T11:36:46.972243+0000 mgr.a (mgr.14152) 190 : audit [DBG] from='client.14376 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:49.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:48 vm09 bash[20661]: audit 2026-03-31T11:36:46.972243+0000 mgr.a (mgr.14152) 190 : audit [DBG] from='client.14376 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:49.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:48 vm09 bash[20661]: cluster 2026-03-31T11:36:47.496513+0000 mgr.a (mgr.14152) 191 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:49.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:48 vm09 bash[20661]: cluster 2026-03-31T11:36:47.496513+0000 mgr.a (mgr.14152) 191 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:49.658 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:49.906 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:49.959 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:49 vm06 bash[20457]: audit 2026-03-31T11:36:48.603867+0000 mgr.a (mgr.14152) 192 : audit [DBG] from='client.14380 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:49.959 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:49 vm06 bash[20457]: audit 2026-03-31T11:36:48.603867+0000 mgr.a (mgr.14152) 192 : audit [DBG] from='client.14380 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:50.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:49 vm09 bash[20661]: audit 2026-03-31T11:36:48.603867+0000 mgr.a (mgr.14152) 192 : audit [DBG] from='client.14380 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:50.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:49 vm09 bash[20661]: audit 2026-03-31T11:36:48.603867+0000 mgr.a (mgr.14152) 192 : audit [DBG] from='client.14380 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:50.225 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:50.225 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:46.740042Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:50.295 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:51.018 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:50 vm06 bash[20457]: cluster 2026-03-31T11:36:49.496782+0000 mgr.a (mgr.14152) 193 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:51.018 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:50 vm06 bash[20457]: cluster 2026-03-31T11:36:49.496782+0000 mgr.a (mgr.14152) 193 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:51.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:50 vm09 bash[20661]: cluster 2026-03-31T11:36:49.496782+0000 mgr.a (mgr.14152) 193 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:51.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:50 vm09 bash[20661]: cluster 2026-03-31T11:36:49.496782+0000 mgr.a (mgr.14152) 193 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:51.295 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:51.549 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:51.926 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:51.926 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:46.740042Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:52.017 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:51 vm06 bash[20457]: audit 2026-03-31T11:36:50.223853+0000 mgr.a (mgr.14152) 194 : audit [DBG] from='client.14384 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:52.017 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:51 vm06 bash[20457]: audit 2026-03-31T11:36:50.223853+0000 mgr.a (mgr.14152) 194 : audit [DBG] from='client.14384 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:52.047 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:52.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:51 vm09 bash[20661]: audit 2026-03-31T11:36:50.223853+0000 mgr.a (mgr.14152) 194 : audit [DBG] from='client.14384 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:52.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:51 vm09 bash[20661]: audit 2026-03-31T11:36:50.223853+0000 mgr.a (mgr.14152) 194 : audit [DBG] from='client.14384 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:53.048 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:53.055 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:52 vm06 bash[20457]: cluster 2026-03-31T11:36:51.497037+0000 mgr.a (mgr.14152) 195 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:53.055 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:52 vm06 bash[20457]: cluster 2026-03-31T11:36:51.497037+0000 mgr.a (mgr.14152) 195 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:53.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:52 vm09 bash[20661]: cluster 2026-03-31T11:36:51.497037+0000 mgr.a (mgr.14152) 195 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:53.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:52 vm09 bash[20661]: cluster 2026-03-31T11:36:51.497037+0000 mgr.a (mgr.14152) 195 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:53.300 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:53.862 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:53 vm06 bash[20457]: audit 2026-03-31T11:36:51.925187+0000 mgr.a (mgr.14152) 196 : audit [DBG] from='client.14388 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:53.862 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:53 vm06 bash[20457]: audit 2026-03-31T11:36:51.925187+0000 mgr.a (mgr.14152) 196 : audit [DBG] from='client.14388 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:53.891 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:53.891 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:46.740042Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:37.157474Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:53.995 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:54.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:53 vm09 bash[20661]: audit 2026-03-31T11:36:51.925187+0000 mgr.a (mgr.14152) 196 : audit [DBG] from='client.14388 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:54.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:53 vm09 bash[20661]: audit 2026-03-31T11:36:51.925187+0000 mgr.a (mgr.14152) 196 : audit [DBG] from='client.14388 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:54.995 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:55.037 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: cluster 2026-03-31T11:36:53.497303+0000 mgr.a (mgr.14152) 197 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:55.037 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: cluster 2026-03-31T11:36:53.497303+0000 mgr.a (mgr.14152) 197 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:55.037 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.164029+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.037 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.164029+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.037 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.169663+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.037 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.169663+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.173455+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.173455+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.177133+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.177133+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.190247+0000 mon.a (mon.0) 508 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.190247+0000 mon.a (mon.0) 508 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.191964+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.191964+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.192446+0000 mon.a (mon.0) 510 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.192446+0000 mon.a (mon.0) 510 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.202518+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.202518+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.215962+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.215962+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.216452+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.216452+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.216851+0000 mon.a (mon.0) 514 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:55.038 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:54 vm06 bash[20457]: audit 2026-03-31T11:36:54.216851+0000 mon.a (mon.0) 514 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: cluster 2026-03-31T11:36:53.497303+0000 mgr.a (mgr.14152) 197 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: cluster 2026-03-31T11:36:53.497303+0000 mgr.a (mgr.14152) 197 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.164029+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.164029+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.169663+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.169663+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.173455+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.173455+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.177133+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.177133+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.190247+0000 mon.a (mon.0) 508 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.190247+0000 mon.a (mon.0) 508 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.191964+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.191964+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.192446+0000 mon.a (mon.0) 510 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.192446+0000 mon.a (mon.0) 510 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.202518+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.202518+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.215962+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.215962+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.216452+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.216452+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.216851+0000 mon.a (mon.0) 514 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:55.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:54 vm09 bash[20661]: audit 2026-03-31T11:36:54.216851+0000 mon.a (mon.0) 514 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:55.270 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:55.270 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: Stopping Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:36:55.270 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:55.271 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:55.280 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:55.295 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[20457]: debug 2026-03-31T11:36:55.268+0000 7fd1a7c6f640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-31T11:36:55.295 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[20457]: debug 2026-03-31T11:36:55.268+0000 7fd1a7c6f640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-31T11:36:55.560 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36572]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-mon-a 2026-03-31T11:36:55.560 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a.service: Deactivated successfully. 2026-03-31T11:36:55.560 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: Stopped Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:36:55.560 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:55.560 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:55.561 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:55.791 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:55.791 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:54.177419Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:36:47.374511Z", "running": 1, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 systemd[1]: Started Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.688+0000 7fd6f5438d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.688+0000 7fd6f5438d80 0 ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable - RelWithDebInfo), process ceph-mon, pid 7 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.688+0000 7fd6f5438d80 0 pidfile_write: ignore empty --pid-file 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 0 load: jerasure load: lrc load: isa 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Git sha 0 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Compile date 2026-03-19 20:43:26 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: DB SUMMARY 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: DB Session ID: LUT9WCKJN22BONAU9S8L 2026-03-31T11:36:55.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: CURRENT file: CURRENT 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: MANIFEST file: MANIFEST-000015 size: 281 Bytes 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000014.log size: 5647846 ; 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.error_if_exists: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.create_if_missing: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.env: 0x55cd883eb4c0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.info_log: 0x55cdabae2000 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.statistics: (nil) 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.use_fsync: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.db_log_dir: 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.wal_dir: 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.write_buffer_manager: 0x55cdabad6140 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.unordered_write: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.row_cache: None 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.wal_filter: None 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-31T11:36:55.817 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.two_write_queues: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.wal_compression: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.atomic_flush: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_open_files: -1 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Compression algorithms supported: 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: kZSTD supported: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: kXpressCompression supported: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: kZlibCompression supported: 1 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-31T11:36:55.818 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.merge_operator: 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_filter: None 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55cdabad2c00) 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cache_index_and_filter_blocks: 1 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: pin_top_level_index_and_filter: 1 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: index_type: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: data_block_index_type: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: index_shortening: 1 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: data_block_hash_table_util_ratio: 0.750000 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: checksum: 4 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: no_block_cache: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: block_cache: 0x55cdabaf5b90 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: block_cache_name: BinnedLRUCache 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: block_cache_options: 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: capacity : 536870912 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: num_shard_bits : 4 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: strict_capacity_limit : 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: high_pri_pool_ratio: 0.000 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: block_cache_compressed: (nil) 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: persistent_cache: (nil) 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: block_size: 4096 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: block_size_deviation: 10 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: block_restart_interval: 16 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: index_block_restart_interval: 1 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: metadata_block_size: 4096 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: partition_filters: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: use_delta_encoding: 1 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: filter_policy: bloomfilter 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: whole_key_filtering: 1 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: verify_compression: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: read_amp_bytes_per_bit: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: format_version: 5 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: enable_index_compression: 1 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: block_align: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: max_auto_readahead_size: 262144 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: prepopulate_block_cache: 0 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: initial_auto_readahead_size: 8192 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: num_file_reads_for_auto_readahead: 2 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-31T11:36:55.819 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression: NoCompression 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.num_levels: 7 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-31T11:36:55.820 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.bloom_locality: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.ttl: 2592000 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.enable_blob_files: false 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.min_blob_size: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.692+0000 7fd6f5438d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.696+0000 7fd6f5438d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 231, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.696+0000 7fd6f5438d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.696+0000 7fd6f5438d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 31fd3944-8735-4c92-b547-5b1b6a207f76 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.696+0000 7fd6f5438d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957015697110, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.696+0000 7fd6f5438d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.712+0000 7fd6f5438d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957015714631, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 4777924, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 236, "largest_seqno": 4405, "table_properties": {"data_size": 4762387, "index_size": 9952, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 4549, "raw_key_size": 39760, "raw_average_key_size": 21, "raw_value_size": 4726988, "raw_average_value_size": 2604, "num_data_blocks": 476, "num_entries": 1815, "num_filter_entries": 1815, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1774957015, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "31fd3944-8735-4c92-b547-5b1b6a207f76", "db_session_id": "LUT9WCKJN22BONAU9S8L", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.712+0000 7fd6f5438d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957015714808, "job": 1, "event": "recovery_finished"} 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.712+0000 7fd6f5438d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55cdabaf6e00 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 4 rocksdb: DB pointer 0x55cdabb40000 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6eb1a9640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6eb1a9640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: ** DB Stats ** 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:36:55.821 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: ** Compaction Stats [default] ** 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: L0 3/0 4.63 MB 0.8 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 379.9 0.01 0.00 1 0.012 0 0 0.0 0.0 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Sum 3/0 4.63 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 379.9 0.01 0.00 1 0.012 0 0 0.0 0.0 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 379.9 0.01 0.00 1 0.012 0 0 0.0 0.0 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: ** Compaction Stats [default] ** 2026-03-31T11:36:55.822 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 379.9 0.01 0.00 1 0.012 0 0 0.0 0.0 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Flush(GB): cumulative 0.004, interval 0.004 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: AddFile(Total Files): cumulative 0, interval 0 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: AddFile(Keys): cumulative 0, interval 0 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Cumulative compaction: 0.00 GB write, 199.51 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Interval compaction: 0.00 GB write, 199.51 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Block cache BinnedLRUCache@0x55cdabaf5b90#7 capacity: 512.00 MB usage: 15.67 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.9e-05 secs_since: 0 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: Block cache entry stats(count,size,portion): FilterBlock(3,5.23 KB,0.000998378%) IndexBlock(3,10.44 KB,0.0019908%) Misc(1,0.00 KB,0%) 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: ** File Read Latency Histogram By Level [default] ** 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] at bind addrs [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 1 mon.a@-1(???) e2 preinit fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 0 mon.a@-1(???).mds e1 new map 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 0 mon.a@-1(???).mds e1 print_map 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: e1 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: btime 2026-03-31T11:33:11:293707+0000 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: legacy client fscid: -1 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: No filesystems configured 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 0 mon.a@-1(???).osd e16 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 0 mon.a@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 0 mon.a@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:55.823 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: debug 2026-03-31T11:36:55.716+0000 7fd6f5438d80 0 mon.a@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:55.876 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.274165+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.274165+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.274214+0000 mon.b (mon.1) 7 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.274214+0000 mon.b (mon.1) 7 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.497535+0000 mgr.a (mgr.14152) 202 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.497535+0000 mgr.a (mgr.14152) 202 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.790061+0000 mgr.a (mgr.14152) 203 : audit [DBG] from='client.24223 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.790061+0000 mgr.a (mgr.14152) 203 : audit [DBG] from='client.24223 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.930122+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.930122+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.932587+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.932587+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935399+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935399+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935406+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935406+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935409+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935409+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935412+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935412+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935414+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935414+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935416+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935416+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935420+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935420+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935422+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935422+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935744+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935744+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935757+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.935757+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.936316+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.936316+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.936438+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: cluster 2026-03-31T11:36:55.936438+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.943098+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.943098+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.946559+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.946559+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:56.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.949807+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.949807+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.952455+0000 mon.a (mon.0) 18 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.952455+0000 mon.a (mon.0) 18 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.953620+0000 mon.b (mon.1) 8 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.953620+0000 mon.b (mon.1) 8 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.954221+0000 mon.b (mon.1) 9 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.954221+0000 mon.b (mon.1) 9 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.954757+0000 mon.b (mon.1) 10 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:55 vm09 bash[20661]: audit 2026-03-31T11:36:55.954757+0000 mon.b (mon.1) 10 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.274165+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.274165+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.274214+0000 mon.b (mon.1) 7 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.274214+0000 mon.b (mon.1) 7 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.497535+0000 mgr.a (mgr.14152) 202 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.497535+0000 mgr.a (mgr.14152) 202 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.790061+0000 mgr.a (mgr.14152) 203 : audit [DBG] from='client.24223 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.790061+0000 mgr.a (mgr.14152) 203 : audit [DBG] from='client.24223 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.930122+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.930122+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.932587+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.932587+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935399+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935399+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935406+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935406+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935409+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935409+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935412+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935412+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935414+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935414+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935416+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935416+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935420+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935420+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935422+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935422+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935744+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935744+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935757+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.935757+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.936316+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.936316+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.936438+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: cluster 2026-03-31T11:36:55.936438+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.943098+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.943098+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.946559+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.946559+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.949807+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.949807+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.952455+0000 mon.a (mon.0) 18 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.952455+0000 mon.a (mon.0) 18 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:56.086 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.953620+0000 mon.b (mon.1) 8 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:56.087 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.953620+0000 mon.b (mon.1) 8 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:56.087 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.954221+0000 mon.b (mon.1) 9 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:56.087 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.954221+0000 mon.b (mon.1) 9 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:56.087 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.954757+0000 mon.b (mon.1) 10 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:56.087 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:55 vm06 bash[36745]: audit 2026-03-31T11:36:55.954757+0000 mon.b (mon.1) 10 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:56.500 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:56.500 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: Stopping Ceph mon.b for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:36:56.500 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:56.500 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:56.777 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[20661]: debug 2026-03-31T11:36:56.530+0000 7f68e53ac640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-31T11:36:56.777 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[20661]: debug 2026-03-31T11:36:56.530+0000 7f68e53ac640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-31T11:36:56.777 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31449]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-mon-b 2026-03-31T11:36:56.777 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.b.service: Deactivated successfully. 2026-03-31T11:36:56.777 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: Stopped Ceph mon.b for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:36:56.876 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:57.084 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:57.084 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 systemd[1]: Started Ceph mon.b for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.934+0000 7f262a55cd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.934+0000 7f262a55cd80 0 ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable - RelWithDebInfo), process ceph-mon, pid 7 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.934+0000 7f262a55cd80 0 pidfile_write: ignore empty --pid-file 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 0 load: jerasure load: lrc load: isa 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Git sha 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Compile date 2026-03-19 20:43:26 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: DB SUMMARY 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: DB Session ID: RKUQ8LKRGH6IIIQ0HZBD 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: CURRENT file: CURRENT 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 1, files: 000008.sst 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000009.log size: 7468685 ; 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.error_if_exists: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.create_if_missing: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.env: 0x56237478c4c0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.info_log: 0x562395544000 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.statistics: (nil) 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.use_fsync: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.db_log_dir: 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.wal_dir: 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-31T11:36:57.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.write_buffer_manager: 0x562395538140 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.unordered_write: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.row_cache: None 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.wal_filter: None 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.two_write_queues: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.wal_compression: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.atomic_flush: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_open_files: -1 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Compression algorithms supported: 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: kZSTD supported: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: kXpressCompression supported: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-31T11:36:57.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: kZlibCompression supported: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.938+0000 7f262a55cd80 4 rocksdb: Options.merge_operator: 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_filter: None 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x562395534c00) 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: cache_index_and_filter_blocks: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: pin_top_level_index_and_filter: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: index_type: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: data_block_index_type: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: index_shortening: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: data_block_hash_table_util_ratio: 0.750000 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: checksum: 4 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: no_block_cache: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: block_cache: 0x562395557b90 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: block_cache_name: BinnedLRUCache 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: block_cache_options: 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: capacity : 536870912 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: num_shard_bits : 4 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: strict_capacity_limit : 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: high_pri_pool_ratio: 0.000 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: block_cache_compressed: (nil) 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: persistent_cache: (nil) 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: block_size: 4096 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: block_size_deviation: 10 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: block_restart_interval: 16 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: index_block_restart_interval: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: metadata_block_size: 4096 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: partition_filters: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: use_delta_encoding: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: filter_policy: bloomfilter 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: whole_key_filtering: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: verify_compression: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: read_amp_bytes_per_bit: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: format_version: 5 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: enable_index_compression: 1 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: block_align: 0 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: max_auto_readahead_size: 262144 2026-03-31T11:36:57.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: prepopulate_block_cache: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: initial_auto_readahead_size: 8192 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: num_file_reads_for_auto_readahead: 2 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression: NoCompression 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.num_levels: 7 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-31T11:36:57.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.bloom_locality: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.ttl: 2592000 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.enable_blob_files: false 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.min_blob_size: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 1364afb3-22dd-4989-a5dd-93c66f3c9a1e 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957016949791, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.942+0000 7f262a55cd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.962+0000 7f262a55cd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957016967433, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 4918531, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 6, "largest_seqno": 4534, "table_properties": {"data_size": 4901959, "index_size": 10411, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 5125, "raw_key_size": 50089, "raw_average_key_size": 24, "raw_value_size": 4860673, "raw_average_value_size": 2387, "num_data_blocks": 496, "num_entries": 2036, "num_filter_entries": 2036, "num_deletions": 4, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1774957016, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "1364afb3-22dd-4989-a5dd-93c66f3c9a1e", "db_session_id": "RKUQ8LKRGH6IIIQ0HZBD", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.962+0000 7f262a55cd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957016967596, "job": 1, "event": "recovery_finished"} 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.962+0000 7f262a55cd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x562395558e00 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 4 rocksdb: DB pointer 0x5623955a2000 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f26202cd640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f26202cd640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: ** DB Stats ** 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: ** Compaction Stats [default] ** 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: L0 2/0 4.69 MB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 503.7 0.01 0.00 1 0.009 0 0 0.0 0.0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Sum 2/0 4.69 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 503.7 0.01 0.00 1 0.009 0 0 0.0 0.0 2026-03-31T11:36:57.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 503.7 0.01 0.00 1 0.009 0 0 0.0 0.0 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: ** Compaction Stats [default] ** 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 503.7 0.01 0.00 1 0.009 0 0 0.0 0.0 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Flush(GB): cumulative 0.005, interval 0.005 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: AddFile(Total Files): cumulative 0, interval 0 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: AddFile(Keys): cumulative 0, interval 0 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Cumulative compaction: 0.00 GB write, 185.11 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Interval compaction: 0.00 GB write, 185.11 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Block cache BinnedLRUCache@0x562395557b90#7 capacity: 512.00 MB usage: 18.33 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.5e-05 secs_since: 0 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: Block cache entry stats(count,size,portion): FilterBlock(2,6.14 KB,0.00117123%) IndexBlock(2,12.19 KB,0.00232458%) Misc(1,0.00 KB,0%) 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: ** File Read Latency Histogram By Level [default] ** 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 0 starting mon.b rank 1 at public addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] at bind addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 1 mon.b@-1(???) e2 preinit fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 0 mon.b@-1(???).mds e1 new map 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 0 mon.b@-1(???).mds e1 print_map 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: e1 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: btime 2026-03-31T11:33:11:293707+0000 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: legacy client fscid: -1 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: No filesystems configured 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 0 mon.b@-1(???).osd e16 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 0 mon.b@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 0 mon.b@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 0 mon.b@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:57.092 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:56 vm09 bash[31577]: debug 2026-03-31T11:36:56.966+0000 7f262a55cd80 1 mon.b@-1(???).paxosservice(auth 1..10) refresh upgraded, format 0 -> 3 2026-03-31T11:36:57.127 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:57.436 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:36:57.437 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:54.177419Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:36:57.517 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:36:57.706 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:36:57 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:57.707 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:36:57 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:57.707 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:57 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:57.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:57 vm06 systemd[1]: Stopping Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:36:57.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:57 vm06 bash[36745]: debug 2026-03-31T11:36:57.756+0000 7fd6f39ba640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-31T11:36:57.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:57 vm06 bash[36745]: debug 2026-03-31T11:36:57.756+0000 7fd6f39ba640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-31T11:36:57.957 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:57 vm06 bash[37252]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-mon-a 2026-03-31T11:36:58.208 INFO:journalctl@ceph.mgr.a.vm06.stdout:Mar 31 11:36:58 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:58.209 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:36:58 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:58.209 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:57 vm06 systemd[1]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a.service: Deactivated successfully. 2026-03-31T11:36:58.209 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:57 vm06 systemd[1]: Stopped Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:36:58.209 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:36:58.209 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 systemd[1]: Started Ceph mon.a for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:36:58.517 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.228+0000 7fc7c3009d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.228+0000 7fc7c3009d80 0 ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable - RelWithDebInfo), process ceph-mon, pid 7 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.228+0000 7fc7c3009d80 0 pidfile_write: ignore empty --pid-file 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 0 load: jerasure load: lrc load: isa 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Git sha 0 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Compile date 2026-03-19 20:43:26 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: DB SUMMARY 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: DB Session ID: H8BV8F4JUIOW7B7DMW25 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: CURRENT file: CURRENT 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-31T11:36:58.523 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: MANIFEST file: MANIFEST-000020 size: 373 Bytes 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 3, files: 000008.sst 000013.sst 000018.sst 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000019.log size: 189403 ; 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.error_if_exists: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.create_if_missing: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.env: 0x562e615f34c0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.info_log: 0x562e92e5c000 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.statistics: (nil) 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.use_fsync: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.db_log_dir: 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.wal_dir: 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.write_buffer_manager: 0x562e92e50140 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.unordered_write: 0 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.row_cache: None 2026-03-31T11:36:58.524 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.wal_filter: None 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.two_write_queues: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.wal_compression: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.atomic_flush: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_open_files: -1 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Compression algorithms supported: 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: kZSTD supported: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: kXpressCompression supported: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: kZlibCompression supported: 1 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.merge_operator: 2026-03-31T11:36:58.525 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_filter: None 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x562e92e4cc00) 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: cache_index_and_filter_blocks: 1 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: pin_top_level_index_and_filter: 1 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: index_type: 0 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: data_block_index_type: 0 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: index_shortening: 1 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: data_block_hash_table_util_ratio: 0.750000 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: checksum: 4 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: no_block_cache: 0 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: block_cache: 0x562e92e6fb90 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: block_cache_name: BinnedLRUCache 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: block_cache_options: 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: capacity : 536870912 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: num_shard_bits : 4 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: strict_capacity_limit : 0 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: high_pri_pool_ratio: 0.000 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: block_cache_compressed: (nil) 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: persistent_cache: (nil) 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: block_size: 4096 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: block_size_deviation: 10 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: block_restart_interval: 16 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: index_block_restart_interval: 1 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: metadata_block_size: 4096 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: partition_filters: 0 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: use_delta_encoding: 1 2026-03-31T11:36:58.526 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: filter_policy: bloomfilter 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: whole_key_filtering: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: verify_compression: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: read_amp_bytes_per_bit: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: format_version: 5 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: enable_index_compression: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: block_align: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: max_auto_readahead_size: 262144 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: prepopulate_block_cache: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: initial_auto_readahead_size: 8192 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: num_file_reads_for_auto_readahead: 2 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression: NoCompression 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.num_levels: 7 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-31T11:36:58.527 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.bloom_locality: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.ttl: 2592000 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.enable_blob_files: false 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.min_blob_size: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.232+0000 7fc7c3009d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.236+0000 7fc7c3009d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 succeeded,manifest_file_number is 20, next_file_number is 22, last_sequence is 4405, log_number is 15,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 15 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.236+0000 7fc7c3009d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 15 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.236+0000 7fc7c3009d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 31fd3944-8735-4c92-b547-5b1b6a207f76 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.236+0000 7fc7c3009d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957018240682, "job": 1, "event": "recovery_started", "wal_files": [19]} 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.236+0000 7fc7c3009d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #19 mode 2 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.240+0000 7fc7c3009d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957018242380, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 23, "file_size": 166716, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 4411, "largest_seqno": 4540, "table_properties": {"data_size": 164907, "index_size": 584, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 197, "raw_key_size": 1870, "raw_average_key_size": 25, "raw_value_size": 163041, "raw_average_value_size": 2233, "num_data_blocks": 23, "num_entries": 73, "num_filter_entries": 73, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1774957018, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "31fd3944-8735-4c92-b547-5b1b6a207f76", "db_session_id": "H8BV8F4JUIOW7B7DMW25", "orig_file_number": 23, "seqno_to_time_mapping": "N/A"}} 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.240+0000 7fc7c3009d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957018242560, "job": 1, "event": "recovery_finished"} 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.240+0000 7fc7c3009d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 25 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000019.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x562e92e70e00 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 4 rocksdb: DB pointer 0x562e92eba000 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c0d8a640 4 rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 3] Compacting 4@0 files to L6, score 1.00 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c0d8a640 4 rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 2 Base level 0, inputs: [23(162KB) 18(4665KB) 13(72KB) 8(1900B)] 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c0d8a640 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957018245106, "job": 3, "event": "compaction_started", "compaction_reason": "LevelL0FilesNum", "files_L0": [23, 18, 13, 8], "score": 1, "input_data_size": 5020587, "oldest_snapshot_seqno": -1} 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7b8d7a640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7b8d7a640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: ** DB Stats ** 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: ** Compaction Stats [default] ** 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:36:58.528 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: L0 4/4 4.79 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 137.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Sum 4/4 4.79 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 137.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 137.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: ** Compaction Stats [default] ** 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 137.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: AddFile(Total Files): cumulative 0, interval 0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: AddFile(Keys): cumulative 0, interval 0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Cumulative compaction: 0.00 GB write, 17.97 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Interval compaction: 0.00 GB write, 17.97 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Block cache BinnedLRUCache@0x562e92e6fb90#7 capacity: 512.00 MB usage: 16.61 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1e-05 secs_since: 0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: Block cache entry stats(count,size,portion): FilterBlock(4,5.47 KB,0.00104308%) IndexBlock(4,11.14 KB,0.00212491%) Misc(1,0.00 KB,0%) 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: ** File Read Latency Histogram By Level [default] ** 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 10 CompressorRegistry(0x562e9313ca18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7b9d7c640 2 Event(0x562e92fc1288 nevent=5000 time_id=1).set_owner center_id=0 owner=140495793145408 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7b9d7c640 10 stack operator() starting 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7bfd88640 2 Event(0x562e92fc0e08 nevent=5000 time_id=1).set_owner center_id=2 owner=140495893857856 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7ba57d640 2 Event(0x562e92fc1048 nevent=5000 time_id=1).set_owner center_id=1 owner=140495801538112 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7ba57d640 10 stack operator() starting 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7bfd88640 10 stack operator() starting 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] at bind addrs [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 10 CompressorRegistry(0x562e9313d218) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 10 -- operator() registering messenger mon-mgrc using key mon-mgrc failed due to a name collision. messenger won't be available to "messenger dump" 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 1 mon.a@-1(???) e2 preinit fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 0 mon.a@-1(???).mds e1 new map 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 0 mon.a@-1(???).mds e1 print_map 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: e1 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: btime 2026-03-31T11:33:11:293707+0000 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: legacy client fscid: -1 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: No filesystems configured 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 0 mon.a@-1(???).osd e16 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 0 mon.a@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 0 mon.a@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 0 mon.a@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:36:58.529 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:58 vm06 bash[37427]: debug 2026-03-31T11:36:58.244+0000 7fc7c3009d80 1 mon.a@-1(???).paxosservice(auth 1..10) refresh upgraded, format 0 -> 3 2026-03-31T11:36:58.774 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cephadm 2026-03-31T11:36:57.116818+0000 mgr.a (mgr.14152) 207 : cephadm [INF] Redeploying mon.a, (entrypoint args changed) . . . 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cephadm 2026-03-31T11:36:57.116818+0000 mgr.a (mgr.14152) 207 : cephadm [INF] Redeploying mon.a, (entrypoint args changed) . . . 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cephadm 2026-03-31T11:36:57.118385+0000 mgr.a (mgr.14152) 208 : cephadm [INF] Deploying daemon mon.a on vm06 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cephadm 2026-03-31T11:36:57.118385+0000 mgr.a (mgr.14152) 208 : cephadm [INF] Deploying daemon mon.a on vm06 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:57.432302+0000 mgr.a (mgr.14152) 209 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:57.432302+0000 mgr.a (mgr.14152) 209 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:57.498065+0000 mgr.a (mgr.14152) 210 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:57.498065+0000 mgr.a (mgr.14152) 210 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:57.765304+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:57.765304+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:57.765407+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:57.765407+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.656258+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.656258+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.658916+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.658916+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662643+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662643+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662662+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662662+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:36:59.945 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662675+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662675+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662684+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662684+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662696+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662696+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662705+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662705+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662718+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662718+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662727+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.662727+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.663083+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.663083+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.663111+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.663111+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.663782+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.663782+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.663898+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: cluster 2026-03-31T11:36:59.663898+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.669190+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.669190+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.671611+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.671611+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.675464+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.675464+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.676754+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.676754+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.677239+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.677239+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.677703+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:36:59.946 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:36:59 vm09 bash[31577]: audit 2026-03-31T11:36:59.677703+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cephadm 2026-03-31T11:36:57.116818+0000 mgr.a (mgr.14152) 207 : cephadm [INF] Redeploying mon.a, (entrypoint args changed) . . . 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cephadm 2026-03-31T11:36:57.116818+0000 mgr.a (mgr.14152) 207 : cephadm [INF] Redeploying mon.a, (entrypoint args changed) . . . 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cephadm 2026-03-31T11:36:57.118385+0000 mgr.a (mgr.14152) 208 : cephadm [INF] Deploying daemon mon.a on vm06 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cephadm 2026-03-31T11:36:57.118385+0000 mgr.a (mgr.14152) 208 : cephadm [INF] Deploying daemon mon.a on vm06 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:57.432302+0000 mgr.a (mgr.14152) 209 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:57.432302+0000 mgr.a (mgr.14152) 209 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:57.498065+0000 mgr.a (mgr.14152) 210 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:57.498065+0000 mgr.a (mgr.14152) 210 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:57.765304+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:57.765304+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:57.765407+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:57.765407+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.656258+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.656258+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.658916+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.658916+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662643+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662643+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662662+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662662+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662675+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662675+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662684+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662684+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662696+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662696+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662705+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662705+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662718+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:37:00.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662718+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662727+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.662727+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.663083+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.663083+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.663111+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.663111+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.663782+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.663782+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.663898+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: cluster 2026-03-31T11:36:59.663898+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.669190+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.669190+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.671611+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.671611+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14152 ' entity='' 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.675464+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.675464+0000 mon.a (mon.0) 17 : audit [INF] from='mgr.14152 ' entity='mgr.a' 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.676754+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.676754+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.677239+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.677239+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.677703+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:00.067 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:36:59 vm06 bash[37427]: audit 2026-03-31T11:36:59.677703+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:00.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:37:00.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: Stopping Ceph mon.b for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:37:00.230 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:37:00.230 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:37:00.484 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[31577]: debug 2026-03-31T11:37:00.262+0000 7f2628ade640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-31T11:37:00.484 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[31577]: debug 2026-03-31T11:37:00.262+0000 7f2628ade640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-31T11:37:00.502 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:37:00.502 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:54.177419Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:37:00.592 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:37:00.756 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[31930]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-mon-b 2026-03-31T11:37:00.756 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.b.service: Deactivated successfully. 2026-03-31T11:37:00.756 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: Stopped Ceph mon.b for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:37:00.756 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:37:00.757 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:37:00.757 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: /etc/systemd/system/ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 systemd[1]: Started Ceph mon.b for 48134446-2cf5-11f1-830d-4121aa8d72d5. 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.854+0000 7fe7b6de6d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.854+0000 7fe7b6de6d80 0 ceph version 20.2.0-721-g5bb32787 (5bb3278730741031382ca9c3dc9d221a942e06a2) tentacle (stable - RelWithDebInfo), process ceph-mon, pid 7 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.854+0000 7fe7b6de6d80 0 pidfile_write: ignore empty --pid-file 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.854+0000 7fe7b6de6d80 0 load: jerasure load: lrc load: isa 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Git sha 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Compile date 2026-03-19 20:43:26 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: DB SUMMARY 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: DB Session ID: XENG5PJ7PADLV05U6YX1 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: CURRENT file: CURRENT 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: MANIFEST file: MANIFEST-000015 size: 282 Bytes 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000014.log size: 244724 ; 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.error_if_exists: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.create_if_missing: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.env: 0x558107ce04c0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.info_log: 0x5581318cc000 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.statistics: (nil) 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.use_fsync: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.db_log_dir: 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.wal_dir: 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-31T11:37:01.086 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.write_buffer_manager: 0x5581318c0140 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.unordered_write: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.row_cache: None 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.wal_filter: None 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.two_write_queues: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.wal_compression: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.atomic_flush: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_open_files: -1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Compression algorithms supported: 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: kZSTD supported: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: kXpressCompression supported: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: kZlibCompression supported: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-31T11:37:01.087 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.merge_operator: 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_filter: None 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5581318bcc00) 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: cache_index_and_filter_blocks: 1 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: pin_top_level_index_and_filter: 1 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: index_type: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: data_block_index_type: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: index_shortening: 1 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: data_block_hash_table_util_ratio: 0.750000 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: checksum: 4 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: no_block_cache: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: block_cache: 0x5581318dfb90 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: block_cache_name: BinnedLRUCache 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: block_cache_options: 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: capacity : 536870912 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: num_shard_bits : 4 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: strict_capacity_limit : 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: high_pri_pool_ratio: 0.000 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: block_cache_compressed: (nil) 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: persistent_cache: (nil) 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: block_size: 4096 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: block_size_deviation: 10 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: block_restart_interval: 16 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: index_block_restart_interval: 1 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: metadata_block_size: 4096 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: partition_filters: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: use_delta_encoding: 1 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: filter_policy: bloomfilter 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: whole_key_filtering: 1 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: verify_compression: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: read_amp_bytes_per_bit: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: format_version: 5 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: enable_index_compression: 1 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: block_align: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: max_auto_readahead_size: 262144 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: prepopulate_block_cache: 0 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: initial_auto_readahead_size: 8192 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: num_file_reads_for_auto_readahead: 2 2026-03-31T11:37:01.088 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression: NoCompression 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.num_levels: 7 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-31T11:37:01.089 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.bloom_locality: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.ttl: 2592000 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.enable_blob_files: false 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.min_blob_size: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 4534, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 1364afb3-22dd-4989-a5dd-93c66f3c9a1e 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957020863376, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957020864796, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 155063, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 4546, "largest_seqno": 4664, "table_properties": {"data_size": 153304, "index_size": 534, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 197, "raw_key_size": 1844, "raw_average_key_size": 25, "raw_value_size": 151481, "raw_average_value_size": 2133, "num_data_blocks": 22, "num_entries": 71, "num_filter_entries": 71, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1774957020, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "1364afb3-22dd-4989-a5dd-93c66f3c9a1e", "db_session_id": "XENG5PJ7PADLV05U6YX1", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774957020864849, "job": 1, "event": "recovery_finished"} 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.858+0000 7fe7b6de6d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5581318e0e00 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 4 rocksdb: DB pointer 0x558131a30000 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7acb57640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7acb57640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: ** DB Stats ** 2026-03-31T11:37:01.090 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: ** Compaction Stats [default] ** 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: L0 3/0 4.84 MB 0.8 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 131.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Sum 3/0 4.84 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 131.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 131.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: ** Compaction Stats [default] ** 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 131.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: AddFile(Total Files): cumulative 0, interval 0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: AddFile(Keys): cumulative 0, interval 0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Cumulative compaction: 0.00 GB write, 15.38 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Interval compaction: 0.00 GB write, 15.38 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Block cache BinnedLRUCache@0x5581318dfb90#7 capacity: 512.00 MB usage: 19.20 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.7e-05 secs_since: 0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: Block cache entry stats(count,size,portion): FilterBlock(3,6.38 KB,0.00121593%) IndexBlock(3,12.83 KB,0.00244677%) Misc(1,0.00 KB,0%) 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: ** File Read Latency Histogram By Level [default] ** 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 10 CompressorRegistry(0x558131baca18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7adb59640 2 Event(0x55813192ae08 nevent=5000 time_id=1).set_owner center_id=0 owner=140633028531776 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7adb59640 10 stack operator() starting 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7ae35a640 2 Event(0x55813192abc8 nevent=5000 time_id=1).set_owner center_id=1 owner=140633036924480 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7ae35a640 10 stack operator() starting 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b3b65640 2 Event(0x558132711708 nevent=5000 time_id=1).set_owner center_id=2 owner=140633129244224 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b3b65640 10 stack operator() starting 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 0 starting mon.b rank 1 at public addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] at bind addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 10 CompressorRegistry(0x558131bad218) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 10 -- operator() registering messenger mon-mgrc using key mon-mgrc failed due to a name collision. messenger won't be available to "messenger dump" 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 1 mon.b@-1(???) e2 preinit fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 0 mon.b@-1(???).mds e1 new map 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 0 mon.b@-1(???).mds e1 print_map 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: e1 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: btime 2026-03-31T11:33:11:293707+0000 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: legacy client fscid: -1 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: No filesystems configured 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 0 mon.b@-1(???).osd e16 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 0 mon.b@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 0 mon.b@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 0 mon.b@-1(???).osd e16 crush map has features 288514050185494528, adjusting msgr requires 2026-03-31T11:37:01.091 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:00 vm09 bash[32072]: debug 2026-03-31T11:37:00.866+0000 7fe7b6de6d80 1 mon.b@-1(???).paxosservice(auth 1..11) refresh upgraded, format 0 -> 3 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: cluster 2026-03-31T11:36:59.498325+0000 mgr.a (mgr.14152) 211 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: cluster 2026-03-31T11:36:59.498325+0000 mgr.a (mgr.14152) 211 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: cephadm 2026-03-31T11:36:59.676648+0000 mgr.a (mgr.14152) 212 : cephadm [INF] Redeploying mon.b, (entrypoint args changed) . . . 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: cephadm 2026-03-31T11:36:59.676648+0000 mgr.a (mgr.14152) 212 : cephadm [INF] Redeploying mon.b, (entrypoint args changed) . . . 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: cephadm 2026-03-31T11:36:59.678391+0000 mgr.a (mgr.14152) 213 : cephadm [INF] Deploying daemon mon.b on vm09 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: cephadm 2026-03-31T11:36:59.678391+0000 mgr.a (mgr.14152) 213 : cephadm [INF] Deploying daemon mon.b on vm09 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: audit 2026-03-31T11:37:00.275266+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: audit 2026-03-31T11:37:00.275266+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: audit 2026-03-31T11:37:00.275492+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:37:01.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:01 vm06 bash[37427]: audit 2026-03-31T11:37:00.275492+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: cluster 2026-03-31T11:36:59.498325+0000 mgr.a (mgr.14152) 211 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: cluster 2026-03-31T11:36:59.498325+0000 mgr.a (mgr.14152) 211 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: cephadm 2026-03-31T11:36:59.676648+0000 mgr.a (mgr.14152) 212 : cephadm [INF] Redeploying mon.b, (entrypoint args changed) . . . 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: cephadm 2026-03-31T11:36:59.676648+0000 mgr.a (mgr.14152) 212 : cephadm [INF] Redeploying mon.b, (entrypoint args changed) . . . 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: cephadm 2026-03-31T11:36:59.678391+0000 mgr.a (mgr.14152) 213 : cephadm [INF] Deploying daemon mon.b on vm09 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: cephadm 2026-03-31T11:36:59.678391+0000 mgr.a (mgr.14152) 213 : cephadm [INF] Deploying daemon mon.b on vm09 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: audit 2026-03-31T11:37:00.275266+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: audit 2026-03-31T11:37:00.275266+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: audit 2026-03-31T11:37:00.275492+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:37:01.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:01 vm09 bash[32072]: audit 2026-03-31T11:37:00.275492+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-31T11:37:01.593 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:37:01.832 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:37:02.161 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:37:02.161 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:54.177419Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:37:02.212 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:00.501065+0000 mgr.a (mgr.14152) 214 : audit [DBG] from='client.44100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:00.501065+0000 mgr.a (mgr.14152) 214 : audit [DBG] from='client.44100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.267290+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.267290+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.269114+0000 mon.a (mon.0) 20 : cluster [INF] mon.a calling monitor election 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.269114+0000 mon.a (mon.0) 20 : cluster [INF] mon.a calling monitor election 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.271094+0000 mon.a (mon.0) 21 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.271094+0000 mon.a (mon.0) 21 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276785+0000 mon.a (mon.0) 22 : cluster [DBG] monmap epoch 2 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276785+0000 mon.a (mon.0) 22 : cluster [DBG] monmap epoch 2 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276887+0000 mon.a (mon.0) 23 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276887+0000 mon.a (mon.0) 23 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276897+0000 mon.a (mon.0) 24 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276897+0000 mon.a (mon.0) 24 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276907+0000 mon.a (mon.0) 25 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276907+0000 mon.a (mon.0) 25 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276942+0000 mon.a (mon.0) 26 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276942+0000 mon.a (mon.0) 26 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276968+0000 mon.a (mon.0) 27 : cluster [DBG] election_strategy: 1 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276968+0000 mon.a (mon.0) 27 : cluster [DBG] election_strategy: 1 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276977+0000 mon.a (mon.0) 28 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276977+0000 mon.a (mon.0) 28 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276986+0000 mon.a (mon.0) 29 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.276986+0000 mon.a (mon.0) 29 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.277356+0000 mon.a (mon.0) 30 : cluster [DBG] fsmap 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.277356+0000 mon.a (mon.0) 30 : cluster [DBG] fsmap 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.277377+0000 mon.a (mon.0) 31 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:37:02.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.277377+0000 mon.a (mon.0) 31 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.277570+0000 mon.a (mon.0) 32 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.277570+0000 mon.a (mon.0) 32 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.277651+0000 mon.a (mon.0) 33 : cluster [INF] overall HEALTH_OK 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.277651+0000 mon.a (mon.0) 33 : cluster [INF] overall HEALTH_OK 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.280574+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.280574+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.283961+0000 mon.a (mon.0) 35 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.283961+0000 mon.a (mon.0) 35 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.285060+0000 mon.a (mon.0) 36 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.285060+0000 mon.a (mon.0) 36 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.286067+0000 mon.a (mon.0) 37 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.286067+0000 mon.a (mon.0) 37 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.286480+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.286480+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.289727+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: audit 2026-03-31T11:37:01.289727+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cephadm 2026-03-31T11:37:01.290049+0000 mgr.a (mgr.14152) 215 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cephadm 2026-03-31T11:37:01.290049+0000 mgr.a (mgr.14152) 215 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.498582+0000 mgr.a (mgr.14152) 216 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:02.567 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:02 vm06 bash[37427]: cluster 2026-03-31T11:37:01.498582+0000 mgr.a (mgr.14152) 216 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:00.501065+0000 mgr.a (mgr.14152) 214 : audit [DBG] from='client.44100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:00.501065+0000 mgr.a (mgr.14152) 214 : audit [DBG] from='client.44100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.267290+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.267290+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.269114+0000 mon.a (mon.0) 20 : cluster [INF] mon.a calling monitor election 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.269114+0000 mon.a (mon.0) 20 : cluster [INF] mon.a calling monitor election 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.271094+0000 mon.a (mon.0) 21 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.271094+0000 mon.a (mon.0) 21 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276785+0000 mon.a (mon.0) 22 : cluster [DBG] monmap epoch 2 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276785+0000 mon.a (mon.0) 22 : cluster [DBG] monmap epoch 2 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276887+0000 mon.a (mon.0) 23 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276887+0000 mon.a (mon.0) 23 : cluster [DBG] fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276897+0000 mon.a (mon.0) 24 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276897+0000 mon.a (mon.0) 24 : cluster [DBG] last_changed 2026-03-31T11:33:47.318700+0000 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276907+0000 mon.a (mon.0) 25 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276907+0000 mon.a (mon.0) 25 : cluster [DBG] created 2026-03-31T11:33:10.045398+0000 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276942+0000 mon.a (mon.0) 26 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276942+0000 mon.a (mon.0) 26 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276968+0000 mon.a (mon.0) 27 : cluster [DBG] election_strategy: 1 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276968+0000 mon.a (mon.0) 27 : cluster [DBG] election_strategy: 1 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276977+0000 mon.a (mon.0) 28 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276977+0000 mon.a (mon.0) 28 : cluster [DBG] 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-31T11:37:02.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276986+0000 mon.a (mon.0) 29 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.276986+0000 mon.a (mon.0) 29 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.277356+0000 mon.a (mon.0) 30 : cluster [DBG] fsmap 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.277356+0000 mon.a (mon.0) 30 : cluster [DBG] fsmap 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.277377+0000 mon.a (mon.0) 31 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.277377+0000 mon.a (mon.0) 31 : cluster [DBG] osdmap e16: 2 total, 2 up, 2 in 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.277570+0000 mon.a (mon.0) 32 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.277570+0000 mon.a (mon.0) 32 : cluster [DBG] mgrmap e13: a(active, since 3m), standbys: b 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.277651+0000 mon.a (mon.0) 33 : cluster [INF] overall HEALTH_OK 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.277651+0000 mon.a (mon.0) 33 : cluster [INF] overall HEALTH_OK 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.280574+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.280574+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.283961+0000 mon.a (mon.0) 35 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.283961+0000 mon.a (mon.0) 35 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.285060+0000 mon.a (mon.0) 36 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.285060+0000 mon.a (mon.0) 36 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.286067+0000 mon.a (mon.0) 37 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.286067+0000 mon.a (mon.0) 37 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.286480+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.286480+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.289727+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: audit 2026-03-31T11:37:01.289727+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cephadm 2026-03-31T11:37:01.290049+0000 mgr.a (mgr.14152) 215 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cephadm 2026-03-31T11:37:01.290049+0000 mgr.a (mgr.14152) 215 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.498582+0000 mgr.a (mgr.14152) 216 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:02.586 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:02 vm09 bash[32072]: cluster 2026-03-31T11:37:01.498582+0000 mgr.a (mgr.14152) 216 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:03.213 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:37:03.453 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:37:03.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:03 vm06 bash[37427]: audit 2026-03-31T11:37:02.159395+0000 mgr.a (mgr.14152) 217 : audit [DBG] from='client.44104 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:03.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:03 vm06 bash[37427]: audit 2026-03-31T11:37:02.159395+0000 mgr.a (mgr.14152) 217 : audit [DBG] from='client.44104 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:03.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:03 vm06 bash[37427]: audit 2026-03-31T11:37:02.870483+0000 mon.a (mon.0) 40 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:03.492 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:03 vm06 bash[37427]: audit 2026-03-31T11:37:02.870483+0000 mon.a (mon.0) 40 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:03.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:03 vm09 bash[32072]: audit 2026-03-31T11:37:02.159395+0000 mgr.a (mgr.14152) 217 : audit [DBG] from='client.44104 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:03.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:03 vm09 bash[32072]: audit 2026-03-31T11:37:02.159395+0000 mgr.a (mgr.14152) 217 : audit [DBG] from='client.44104 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:03.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:03 vm09 bash[32072]: audit 2026-03-31T11:37:02.870483+0000 mon.a (mon.0) 40 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:03.585 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:03 vm09 bash[32072]: audit 2026-03-31T11:37:02.870483+0000 mon.a (mon.0) 40 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:03.736 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:37:03.736 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:36:37.157541Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:54.177419Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:36:47.374521Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:36:37.157423Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:36:37.157512Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "running": 0, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:36:37.157389Z", "running": 2, "size": 2}}] 2026-03-31T11:37:03.785 INFO:tasks.cephadm:container.bar has 0/2 2026-03-31T11:37:04.565 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:04 vm06 bash[37427]: cluster 2026-03-31T11:37:03.498807+0000 mgr.a (mgr.14152) 218 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:04.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:04 vm06 bash[37427]: cluster 2026-03-31T11:37:03.498807+0000 mgr.a (mgr.14152) 218 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:04.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:04 vm06 bash[37427]: audit 2026-03-31T11:37:03.735047+0000 mgr.a (mgr.14152) 219 : audit [DBG] from='client.44108 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:04.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:04 vm06 bash[37427]: audit 2026-03-31T11:37:03.735047+0000 mgr.a (mgr.14152) 219 : audit [DBG] from='client.44108 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:04.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:04 vm09 bash[32072]: cluster 2026-03-31T11:37:03.498807+0000 mgr.a (mgr.14152) 218 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:04.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:04 vm09 bash[32072]: cluster 2026-03-31T11:37:03.498807+0000 mgr.a (mgr.14152) 218 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:04.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:04 vm09 bash[32072]: audit 2026-03-31T11:37:03.735047+0000 mgr.a (mgr.14152) 219 : audit [DBG] from='client.44108 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:04.584 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:04 vm09 bash[32072]: audit 2026-03-31T11:37:03.735047+0000 mgr.a (mgr.14152) 219 : audit [DBG] from='client.44108 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:04.786 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph orch ls -f json 2026-03-31T11:37:05.042 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:37:05.324 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-31T11:37:05.325 INFO:teuthology.orchestra.run.vm06.stdout:[{"events": ["2026-03-31T11:33:44.749781Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-31T11:33:34.566217Z", "last_refresh": "2026-03-31T11:37:04.761728Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-31T11:36:54.177419Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-31T11:36:28.502358Z", "last_refresh": "2026-03-31T11:37:04.761768Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-31T11:36:29.876219Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-31T11:36:28.494629Z", "last_refresh": "2026-03-31T11:37:04.761592Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:33:55.333038Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06=a", "vm09=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-31T11:33:54.383160Z", "last_refresh": "2026-03-31T11:37:04.761701Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:36:28.508600Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-31T11:36:28.490155Z", "last_refresh": "2026-03-31T11:37:04.761658Z", "running": 2, "size": 2}}, {"events": ["2026-03-31T11:34:05.921263Z service:osd.default [INFO] \"service was created\""], "placement": {"host_pattern": "vm09"}, "service_id": "default", "service_name": "osd.default", "service_type": "osd", "spec": {"data_devices": {"paths": ["vg_nvme/lv_4"]}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-31T11:34:05.917917Z", "last_refresh": "2026-03-31T11:37:04.761535Z", "running": 2, "size": 2}}] 2026-03-31T11:37:05.371 INFO:tasks.cephadm:container.bar has 2/2 2026-03-31T11:37:05.372 INFO:teuthology.run_tasks:Running task exec... 2026-03-31T11:37:05.374 INFO:teuthology.task.exec:Executing custom commands... 2026-03-31T11:37:05.374 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm06.local 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'set -ex 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> sleep 60 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # check extra container and entrypoint args written to mon unit run file 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # check that custom container properly wrote content to file. 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # This requires the custom config, extra container args, and 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # entrypoint args to all be working in order for this to have 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # been written. The container entrypoint was set up with custom_configs, 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # the content and where to write to with the entrypoint args, and the mounting 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # of the /etc/cephadm_testing dir with extra container args 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> grep "testing_custom_containers" /etc/cephadm_testing/testing.txt 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # Verify that container bar'"'"'s init containers and primary container 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> # ran successfully 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name '"'"'container.bar.*'"'"') 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> test -n "$dir" 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> grep ok ${dir}/data/primary.txt 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> grep from=centos ${dir}/data/from.txt 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> test -s ${dir}/data/presized.dat 2026-03-31T11:37:05.374 DEBUG:teuthology.orchestra.run.vm06:> ' 2026-03-31T11:37:05.381 INFO:teuthology.orchestra.run.vm06.stderr:++ /home/ubuntu/cephtest/cephadm shell -- ceph fsid 2026-03-31T11:37:05.615 INFO:teuthology.orchestra.run.vm06.stderr:Inferring fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:05.615 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:37:05.672 INFO:teuthology.orchestra.run.vm06.stderr:Not using image 'sha256:1e58a3cbf9abfa7cd4c97d6122dfc897574d910096f68804997a3e0f45bc44f0' of ceph daemon: not in the list of non-dangling images with ceph=True label 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.766682+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.766682+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.770346+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.770346+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.774312+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.774312+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.879190+0000 mon.a (mon.0) 44 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.879190+0000 mon.a (mon.0) 44 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.882658+0000 mon.a (mon.0) 45 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.882658+0000 mon.a (mon.0) 45 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.884349+0000 mon.a (mon.0) 46 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.884349+0000 mon.a (mon.0) 46 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.886439+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.886439+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.889192+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.889192+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.889692+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.889692+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.892733+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.892733+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.894052+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:37:06.014 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:05 vm06 bash[37427]: audit 2026-03-31T11:37:04.894052+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:37:06.051 INFO:teuthology.orchestra.run.vm06.stderr:+ FSID=48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:37:06.051 INFO:teuthology.orchestra.run.vm06.stderr:+ sleep 60 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.766682+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.766682+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.770346+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.770346+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.774312+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.774312+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.879190+0000 mon.a (mon.0) 44 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.879190+0000 mon.a (mon.0) 44 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.882658+0000 mon.a (mon.0) 45 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.882658+0000 mon.a (mon.0) 45 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.884349+0000 mon.a (mon.0) 46 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.884349+0000 mon.a (mon.0) 46 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.886439+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.886439+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.889192+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.889192+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.889692+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.889692+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.892733+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.892733+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.894052+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:37:06.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:05 vm09 bash[32072]: audit 2026-03-31T11:37:04.894052+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:37:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:06 vm06 bash[37427]: audit 2026-03-31T11:37:05.323526+0000 mgr.a (mgr.14152) 220 : audit [DBG] from='client.44112 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:06 vm06 bash[37427]: audit 2026-03-31T11:37:05.323526+0000 mgr.a (mgr.14152) 220 : audit [DBG] from='client.44112 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:06 vm06 bash[37427]: cluster 2026-03-31T11:37:05.499067+0000 mgr.a (mgr.14152) 221 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:06 vm06 bash[37427]: cluster 2026-03-31T11:37:05.499067+0000 mgr.a (mgr.14152) 221 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:06 vm06 bash[37427]: audit 2026-03-31T11:37:06.004059+0000 mon.a (mon.0) 52 : audit [DBG] from='client.? 192.168.123.106:0/4137477300' entity='client.admin' cmd={"prefix": "fsid"} : dispatch 2026-03-31T11:37:07.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:06 vm06 bash[37427]: audit 2026-03-31T11:37:06.004059+0000 mon.a (mon.0) 52 : audit [DBG] from='client.? 192.168.123.106:0/4137477300' entity='client.admin' cmd={"prefix": "fsid"} : dispatch 2026-03-31T11:37:07.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:06 vm09 bash[32072]: audit 2026-03-31T11:37:05.323526+0000 mgr.a (mgr.14152) 220 : audit [DBG] from='client.44112 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:07.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:06 vm09 bash[32072]: audit 2026-03-31T11:37:05.323526+0000 mgr.a (mgr.14152) 220 : audit [DBG] from='client.44112 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-31T11:37:07.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:06 vm09 bash[32072]: cluster 2026-03-31T11:37:05.499067+0000 mgr.a (mgr.14152) 221 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:07.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:06 vm09 bash[32072]: cluster 2026-03-31T11:37:05.499067+0000 mgr.a (mgr.14152) 221 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:07.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:06 vm09 bash[32072]: audit 2026-03-31T11:37:06.004059+0000 mon.a (mon.0) 52 : audit [DBG] from='client.? 192.168.123.106:0/4137477300' entity='client.admin' cmd={"prefix": "fsid"} : dispatch 2026-03-31T11:37:07.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:06 vm09 bash[32072]: audit 2026-03-31T11:37:06.004059+0000 mon.a (mon.0) 52 : audit [DBG] from='client.? 192.168.123.106:0/4137477300' entity='client.admin' cmd={"prefix": "fsid"} : dispatch 2026-03-31T11:37:09.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:08 vm06 bash[37427]: cluster 2026-03-31T11:37:07.499328+0000 mgr.a (mgr.14152) 222 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:09.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:08 vm06 bash[37427]: cluster 2026-03-31T11:37:07.499328+0000 mgr.a (mgr.14152) 222 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:09.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:08 vm09 bash[32072]: cluster 2026-03-31T11:37:07.499328+0000 mgr.a (mgr.14152) 222 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:09.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:08 vm09 bash[32072]: cluster 2026-03-31T11:37:07.499328+0000 mgr.a (mgr.14152) 222 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:11.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:10 vm06 bash[37427]: cluster 2026-03-31T11:37:09.499570+0000 mgr.a (mgr.14152) 223 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:11.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:10 vm06 bash[37427]: cluster 2026-03-31T11:37:09.499570+0000 mgr.a (mgr.14152) 223 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:11.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:10 vm09 bash[32072]: cluster 2026-03-31T11:37:09.499570+0000 mgr.a (mgr.14152) 223 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:11.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:10 vm09 bash[32072]: cluster 2026-03-31T11:37:09.499570+0000 mgr.a (mgr.14152) 223 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:13.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:12 vm06 bash[37427]: cluster 2026-03-31T11:37:11.499795+0000 mgr.a (mgr.14152) 224 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:13.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:12 vm06 bash[37427]: cluster 2026-03-31T11:37:11.499795+0000 mgr.a (mgr.14152) 224 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:13.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:12 vm09 bash[32072]: cluster 2026-03-31T11:37:11.499795+0000 mgr.a (mgr.14152) 224 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:13.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:12 vm09 bash[32072]: cluster 2026-03-31T11:37:11.499795+0000 mgr.a (mgr.14152) 224 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:15.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:14 vm06 bash[37427]: cluster 2026-03-31T11:37:13.500036+0000 mgr.a (mgr.14152) 225 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:15.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:14 vm06 bash[37427]: cluster 2026-03-31T11:37:13.500036+0000 mgr.a (mgr.14152) 225 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:15.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:14 vm09 bash[32072]: cluster 2026-03-31T11:37:13.500036+0000 mgr.a (mgr.14152) 225 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:15.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:14 vm09 bash[32072]: cluster 2026-03-31T11:37:13.500036+0000 mgr.a (mgr.14152) 225 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:17.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:16 vm06 bash[37427]: cluster 2026-03-31T11:37:15.500289+0000 mgr.a (mgr.14152) 226 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:17.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:16 vm06 bash[37427]: cluster 2026-03-31T11:37:15.500289+0000 mgr.a (mgr.14152) 226 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:17.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:16 vm09 bash[32072]: cluster 2026-03-31T11:37:15.500289+0000 mgr.a (mgr.14152) 226 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:17.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:16 vm09 bash[32072]: cluster 2026-03-31T11:37:15.500289+0000 mgr.a (mgr.14152) 226 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:19.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:18 vm06 bash[37427]: cluster 2026-03-31T11:37:17.500525+0000 mgr.a (mgr.14152) 227 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:19.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:18 vm06 bash[37427]: cluster 2026-03-31T11:37:17.500525+0000 mgr.a (mgr.14152) 227 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:19.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:18 vm09 bash[32072]: cluster 2026-03-31T11:37:17.500525+0000 mgr.a (mgr.14152) 227 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:19.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:18 vm09 bash[32072]: cluster 2026-03-31T11:37:17.500525+0000 mgr.a (mgr.14152) 227 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:21.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:20 vm06 bash[37427]: cluster 2026-03-31T11:37:19.500767+0000 mgr.a (mgr.14152) 228 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:21.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:20 vm06 bash[37427]: cluster 2026-03-31T11:37:19.500767+0000 mgr.a (mgr.14152) 228 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:21.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:20 vm09 bash[32072]: cluster 2026-03-31T11:37:19.500767+0000 mgr.a (mgr.14152) 228 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:21.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:20 vm09 bash[32072]: cluster 2026-03-31T11:37:19.500767+0000 mgr.a (mgr.14152) 228 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:22.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:22 vm06 bash[37427]: audit 2026-03-31T11:37:21.437215+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:22 vm06 bash[37427]: audit 2026-03-31T11:37:21.437215+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:22 vm06 bash[37427]: audit 2026-03-31T11:37:21.440782+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:22 vm06 bash[37427]: audit 2026-03-31T11:37:21.440782+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:22 vm06 bash[37427]: audit 2026-03-31T11:37:21.444417+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:22 vm06 bash[37427]: audit 2026-03-31T11:37:21.444417+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:22 vm06 bash[37427]: cluster 2026-03-31T11:37:21.501037+0000 mgr.a (mgr.14152) 229 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:22.566 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:22 vm06 bash[37427]: cluster 2026-03-31T11:37:21.501037+0000 mgr.a (mgr.14152) 229 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:22.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:22 vm09 bash[32072]: audit 2026-03-31T11:37:21.437215+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:22 vm09 bash[32072]: audit 2026-03-31T11:37:21.437215+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:22 vm09 bash[32072]: audit 2026-03-31T11:37:21.440782+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:22 vm09 bash[32072]: audit 2026-03-31T11:37:21.440782+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:22 vm09 bash[32072]: audit 2026-03-31T11:37:21.444417+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:22 vm09 bash[32072]: audit 2026-03-31T11:37:21.444417+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:22.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:22 vm09 bash[32072]: cluster 2026-03-31T11:37:21.501037+0000 mgr.a (mgr.14152) 229 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:22.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:22 vm09 bash[32072]: cluster 2026-03-31T11:37:21.501037+0000 mgr.a (mgr.14152) 229 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:24.815 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:24 vm06 bash[37427]: cluster 2026-03-31T11:37:23.501323+0000 mgr.a (mgr.14152) 230 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:24.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:24 vm06 bash[37427]: cluster 2026-03-31T11:37:23.501323+0000 mgr.a (mgr.14152) 230 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:24.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:24 vm09 bash[32072]: cluster 2026-03-31T11:37:23.501323+0000 mgr.a (mgr.14152) 230 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:24.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:24 vm09 bash[32072]: cluster 2026-03-31T11:37:23.501323+0000 mgr.a (mgr.14152) 230 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:26.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:25 vm06 bash[37427]: audit 2026-03-31T11:37:24.849025+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:25 vm06 bash[37427]: audit 2026-03-31T11:37:24.849025+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:25 vm06 bash[37427]: audit 2026-03-31T11:37:24.852289+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:25 vm06 bash[37427]: audit 2026-03-31T11:37:24.852289+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:25 vm06 bash[37427]: audit 2026-03-31T11:37:24.856310+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:25 vm06 bash[37427]: audit 2026-03-31T11:37:24.856310+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:25 vm09 bash[32072]: audit 2026-03-31T11:37:24.849025+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:25 vm09 bash[32072]: audit 2026-03-31T11:37:24.849025+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:25 vm09 bash[32072]: audit 2026-03-31T11:37:24.852289+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:25 vm09 bash[32072]: audit 2026-03-31T11:37:24.852289+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:25 vm09 bash[32072]: audit 2026-03-31T11:37:24.856310+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:26.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:25 vm09 bash[32072]: audit 2026-03-31T11:37:24.856310+0000 mon.a (mon.0) 58 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:27.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:26 vm06 bash[37427]: cluster 2026-03-31T11:37:25.501572+0000 mgr.a (mgr.14152) 231 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:27.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:26 vm06 bash[37427]: cluster 2026-03-31T11:37:25.501572+0000 mgr.a (mgr.14152) 231 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:27.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:26 vm09 bash[32072]: cluster 2026-03-31T11:37:25.501572+0000 mgr.a (mgr.14152) 231 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:27.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:26 vm09 bash[32072]: cluster 2026-03-31T11:37:25.501572+0000 mgr.a (mgr.14152) 231 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:28.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:27 vm06 bash[37427]: cluster 2026-03-31T11:37:27.501814+0000 mgr.a (mgr.14152) 232 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:28.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:27 vm06 bash[37427]: cluster 2026-03-31T11:37:27.501814+0000 mgr.a (mgr.14152) 232 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:28.334 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:27 vm09 bash[32072]: cluster 2026-03-31T11:37:27.501814+0000 mgr.a (mgr.14152) 232 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:28.335 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:27 vm09 bash[32072]: cluster 2026-03-31T11:37:27.501814+0000 mgr.a (mgr.14152) 232 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:30.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:30 vm06 bash[37427]: cluster 2026-03-31T11:37:29.502057+0000 mgr.a (mgr.14152) 233 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:30.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:30 vm06 bash[37427]: cluster 2026-03-31T11:37:29.502057+0000 mgr.a (mgr.14152) 233 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:30.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:30 vm09 bash[32072]: cluster 2026-03-31T11:37:29.502057+0000 mgr.a (mgr.14152) 233 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:30.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:30 vm09 bash[32072]: cluster 2026-03-31T11:37:29.502057+0000 mgr.a (mgr.14152) 233 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:32.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:32 vm09 bash[32072]: cluster 2026-03-31T11:37:31.502260+0000 mgr.a (mgr.14152) 234 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:32.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:32 vm09 bash[32072]: cluster 2026-03-31T11:37:31.502260+0000 mgr.a (mgr.14152) 234 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:33.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:32 vm06 bash[37427]: cluster 2026-03-31T11:37:31.502260+0000 mgr.a (mgr.14152) 234 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:33.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:32 vm06 bash[37427]: cluster 2026-03-31T11:37:31.502260+0000 mgr.a (mgr.14152) 234 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:34.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:34 vm09 bash[32072]: cluster 2026-03-31T11:37:33.502449+0000 mgr.a (mgr.14152) 235 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:34.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:34 vm09 bash[32072]: cluster 2026-03-31T11:37:33.502449+0000 mgr.a (mgr.14152) 235 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:34.836 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:34 vm06 bash[37427]: cluster 2026-03-31T11:37:33.502449+0000 mgr.a (mgr.14152) 235 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:34.836 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:34 vm06 bash[37427]: cluster 2026-03-31T11:37:33.502449+0000 mgr.a (mgr.14152) 235 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:36.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:36 vm09 bash[32072]: cluster 2026-03-31T11:37:35.502687+0000 mgr.a (mgr.14152) 236 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:36.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:36 vm09 bash[32072]: cluster 2026-03-31T11:37:35.502687+0000 mgr.a (mgr.14152) 236 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:37.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:36 vm06 bash[37427]: cluster 2026-03-31T11:37:35.502687+0000 mgr.a (mgr.14152) 236 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:37.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:36 vm06 bash[37427]: cluster 2026-03-31T11:37:35.502687+0000 mgr.a (mgr.14152) 236 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:38.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:38 vm09 bash[32072]: cluster 2026-03-31T11:37:37.502909+0000 mgr.a (mgr.14152) 237 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:38.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:38 vm09 bash[32072]: cluster 2026-03-31T11:37:37.502909+0000 mgr.a (mgr.14152) 237 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:39.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:38 vm06 bash[37427]: cluster 2026-03-31T11:37:37.502909+0000 mgr.a (mgr.14152) 237 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:39.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:38 vm06 bash[37427]: cluster 2026-03-31T11:37:37.502909+0000 mgr.a (mgr.14152) 237 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:41.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:40 vm06 bash[37427]: cluster 2026-03-31T11:37:39.503152+0000 mgr.a (mgr.14152) 238 : cluster [DBG] pgmap v142: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:41.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:40 vm06 bash[37427]: cluster 2026-03-31T11:37:39.503152+0000 mgr.a (mgr.14152) 238 : cluster [DBG] pgmap v142: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:41.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:40 vm09 bash[32072]: cluster 2026-03-31T11:37:39.503152+0000 mgr.a (mgr.14152) 238 : cluster [DBG] pgmap v142: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:41.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:40 vm09 bash[32072]: cluster 2026-03-31T11:37:39.503152+0000 mgr.a (mgr.14152) 238 : cluster [DBG] pgmap v142: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:42 vm06 bash[37427]: cluster 2026-03-31T11:37:41.503468+0000 mgr.a (mgr.14152) 239 : cluster [DBG] pgmap v143: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:42 vm06 bash[37427]: cluster 2026-03-31T11:37:41.503468+0000 mgr.a (mgr.14152) 239 : cluster [DBG] pgmap v143: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:42 vm06 bash[37427]: audit 2026-03-31T11:37:41.519339+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:42 vm06 bash[37427]: audit 2026-03-31T11:37:41.519339+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:42 vm06 bash[37427]: audit 2026-03-31T11:37:41.523350+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:42 vm06 bash[37427]: audit 2026-03-31T11:37:41.523350+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:42 vm06 bash[37427]: audit 2026-03-31T11:37:41.547441+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:42 vm06 bash[37427]: audit 2026-03-31T11:37:41.547441+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:42 vm09 bash[32072]: cluster 2026-03-31T11:37:41.503468+0000 mgr.a (mgr.14152) 239 : cluster [DBG] pgmap v143: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:42.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:42 vm09 bash[32072]: cluster 2026-03-31T11:37:41.503468+0000 mgr.a (mgr.14152) 239 : cluster [DBG] pgmap v143: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:42.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:42 vm09 bash[32072]: audit 2026-03-31T11:37:41.519339+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:42 vm09 bash[32072]: audit 2026-03-31T11:37:41.519339+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:42 vm09 bash[32072]: audit 2026-03-31T11:37:41.523350+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:42 vm09 bash[32072]: audit 2026-03-31T11:37:41.523350+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:42 vm09 bash[32072]: audit 2026-03-31T11:37:41.547441+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:42.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:42 vm09 bash[32072]: audit 2026-03-31T11:37:41.547441+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:44.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:44 vm06 bash[37427]: cluster 2026-03-31T11:37:43.503732+0000 mgr.a (mgr.14152) 240 : cluster [DBG] pgmap v144: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:44.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:44 vm06 bash[37427]: cluster 2026-03-31T11:37:43.503732+0000 mgr.a (mgr.14152) 240 : cluster [DBG] pgmap v144: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:44.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:44 vm09 bash[32072]: cluster 2026-03-31T11:37:43.503732+0000 mgr.a (mgr.14152) 240 : cluster [DBG] pgmap v144: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:44.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:44 vm09 bash[32072]: cluster 2026-03-31T11:37:43.503732+0000 mgr.a (mgr.14152) 240 : cluster [DBG] pgmap v144: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:46.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:45 vm09 bash[32072]: audit 2026-03-31T11:37:44.930388+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:45 vm09 bash[32072]: audit 2026-03-31T11:37:44.930388+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:45 vm09 bash[32072]: audit 2026-03-31T11:37:44.933877+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:45 vm09 bash[32072]: audit 2026-03-31T11:37:44.933877+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:45 vm09 bash[32072]: audit 2026-03-31T11:37:44.937685+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:45 vm09 bash[32072]: audit 2026-03-31T11:37:44.937685+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:45 vm09 bash[32072]: cluster 2026-03-31T11:37:45.504007+0000 mgr.a (mgr.14152) 241 : cluster [DBG] pgmap v145: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:46.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:45 vm09 bash[32072]: cluster 2026-03-31T11:37:45.504007+0000 mgr.a (mgr.14152) 241 : cluster [DBG] pgmap v145: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:46.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:45 vm06 bash[37427]: audit 2026-03-31T11:37:44.930388+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:45 vm06 bash[37427]: audit 2026-03-31T11:37:44.930388+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:45 vm06 bash[37427]: audit 2026-03-31T11:37:44.933877+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.315 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:45 vm06 bash[37427]: audit 2026-03-31T11:37:44.933877+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:45 vm06 bash[37427]: audit 2026-03-31T11:37:44.937685+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:45 vm06 bash[37427]: audit 2026-03-31T11:37:44.937685+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:37:46.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:45 vm06 bash[37427]: cluster 2026-03-31T11:37:45.504007+0000 mgr.a (mgr.14152) 241 : cluster [DBG] pgmap v145: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:46.316 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:45 vm06 bash[37427]: cluster 2026-03-31T11:37:45.504007+0000 mgr.a (mgr.14152) 241 : cluster [DBG] pgmap v145: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:48.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:48 vm06 bash[37427]: cluster 2026-03-31T11:37:47.504230+0000 mgr.a (mgr.14152) 242 : cluster [DBG] pgmap v146: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:48.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:48 vm06 bash[37427]: cluster 2026-03-31T11:37:47.504230+0000 mgr.a (mgr.14152) 242 : cluster [DBG] pgmap v146: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:48.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:48 vm09 bash[32072]: cluster 2026-03-31T11:37:47.504230+0000 mgr.a (mgr.14152) 242 : cluster [DBG] pgmap v146: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:48.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:48 vm09 bash[32072]: cluster 2026-03-31T11:37:47.504230+0000 mgr.a (mgr.14152) 242 : cluster [DBG] pgmap v146: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:50.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:50 vm06 bash[37427]: cluster 2026-03-31T11:37:49.504430+0000 mgr.a (mgr.14152) 243 : cluster [DBG] pgmap v147: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:50.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:50 vm06 bash[37427]: cluster 2026-03-31T11:37:49.504430+0000 mgr.a (mgr.14152) 243 : cluster [DBG] pgmap v147: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:50.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:50 vm09 bash[32072]: cluster 2026-03-31T11:37:49.504430+0000 mgr.a (mgr.14152) 243 : cluster [DBG] pgmap v147: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:50.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:50 vm09 bash[32072]: cluster 2026-03-31T11:37:49.504430+0000 mgr.a (mgr.14152) 243 : cluster [DBG] pgmap v147: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:52.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:52 vm09 bash[32072]: cluster 2026-03-31T11:37:51.504621+0000 mgr.a (mgr.14152) 244 : cluster [DBG] pgmap v148: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:52.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:52 vm09 bash[32072]: cluster 2026-03-31T11:37:51.504621+0000 mgr.a (mgr.14152) 244 : cluster [DBG] pgmap v148: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:53.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:52 vm06 bash[37427]: cluster 2026-03-31T11:37:51.504621+0000 mgr.a (mgr.14152) 244 : cluster [DBG] pgmap v148: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:53.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:52 vm06 bash[37427]: cluster 2026-03-31T11:37:51.504621+0000 mgr.a (mgr.14152) 244 : cluster [DBG] pgmap v148: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:54.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:54 vm09 bash[32072]: cluster 2026-03-31T11:37:53.504839+0000 mgr.a (mgr.14152) 245 : cluster [DBG] pgmap v149: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:54.834 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:54 vm09 bash[32072]: cluster 2026-03-31T11:37:53.504839+0000 mgr.a (mgr.14152) 245 : cluster [DBG] pgmap v149: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:55.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:54 vm06 bash[37427]: cluster 2026-03-31T11:37:53.504839+0000 mgr.a (mgr.14152) 245 : cluster [DBG] pgmap v149: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:55.065 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:54 vm06 bash[37427]: cluster 2026-03-31T11:37:53.504839+0000 mgr.a (mgr.14152) 245 : cluster [DBG] pgmap v149: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:56.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:56 vm06 bash[37427]: cluster 2026-03-31T11:37:55.505029+0000 mgr.a (mgr.14152) 246 : cluster [DBG] pgmap v150: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:56.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:56 vm06 bash[37427]: cluster 2026-03-31T11:37:55.505029+0000 mgr.a (mgr.14152) 246 : cluster [DBG] pgmap v150: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:56.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:56 vm09 bash[32072]: cluster 2026-03-31T11:37:55.505029+0000 mgr.a (mgr.14152) 246 : cluster [DBG] pgmap v150: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:56.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:56 vm09 bash[32072]: cluster 2026-03-31T11:37:55.505029+0000 mgr.a (mgr.14152) 246 : cluster [DBG] pgmap v150: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:58.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:58 vm09 bash[32072]: cluster 2026-03-31T11:37:57.505284+0000 mgr.a (mgr.14152) 247 : cluster [DBG] pgmap v151: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:58.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:37:58 vm09 bash[32072]: cluster 2026-03-31T11:37:57.505284+0000 mgr.a (mgr.14152) 247 : cluster [DBG] pgmap v151: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:59.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:58 vm06 bash[37427]: cluster 2026-03-31T11:37:57.505284+0000 mgr.a (mgr.14152) 247 : cluster [DBG] pgmap v151: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:37:59.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:37:58 vm06 bash[37427]: cluster 2026-03-31T11:37:57.505284+0000 mgr.a (mgr.14152) 247 : cluster [DBG] pgmap v151: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:01.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:00 vm06 bash[37427]: cluster 2026-03-31T11:37:59.505553+0000 mgr.a (mgr.14152) 248 : cluster [DBG] pgmap v152: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:01.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:00 vm06 bash[37427]: cluster 2026-03-31T11:37:59.505553+0000 mgr.a (mgr.14152) 248 : cluster [DBG] pgmap v152: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:01.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:00 vm09 bash[32072]: cluster 2026-03-31T11:37:59.505553+0000 mgr.a (mgr.14152) 248 : cluster [DBG] pgmap v152: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:01.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:00 vm09 bash[32072]: cluster 2026-03-31T11:37:59.505553+0000 mgr.a (mgr.14152) 248 : cluster [DBG] pgmap v152: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:02 vm06 bash[37427]: cluster 2026-03-31T11:38:01.505790+0000 mgr.a (mgr.14152) 249 : cluster [DBG] pgmap v153: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:02 vm06 bash[37427]: cluster 2026-03-31T11:38:01.505790+0000 mgr.a (mgr.14152) 249 : cluster [DBG] pgmap v153: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:02 vm06 bash[37427]: audit 2026-03-31T11:38:01.620144+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:02 vm06 bash[37427]: audit 2026-03-31T11:38:01.620144+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:02 vm06 bash[37427]: audit 2026-03-31T11:38:01.623556+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:02 vm06 bash[37427]: audit 2026-03-31T11:38:01.623556+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:02 vm06 bash[37427]: audit 2026-03-31T11:38:01.626729+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.066 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:02 vm06 bash[37427]: audit 2026-03-31T11:38:01.626729+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:02 vm09 bash[32072]: cluster 2026-03-31T11:38:01.505790+0000 mgr.a (mgr.14152) 249 : cluster [DBG] pgmap v153: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:03.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:02 vm09 bash[32072]: cluster 2026-03-31T11:38:01.505790+0000 mgr.a (mgr.14152) 249 : cluster [DBG] pgmap v153: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:03.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:02 vm09 bash[32072]: audit 2026-03-31T11:38:01.620144+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:02 vm09 bash[32072]: audit 2026-03-31T11:38:01.620144+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:02 vm09 bash[32072]: audit 2026-03-31T11:38:01.623556+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:02 vm09 bash[32072]: audit 2026-03-31T11:38:01.623556+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:02 vm09 bash[32072]: audit 2026-03-31T11:38:01.626729+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:03.085 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:02 vm09 bash[32072]: audit 2026-03-31T11:38:01.626729+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:05.015 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:04 vm06 bash[37427]: cluster 2026-03-31T11:38:03.506007+0000 mgr.a (mgr.14152) 250 : cluster [DBG] pgmap v154: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:05.015 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:04 vm06 bash[37427]: cluster 2026-03-31T11:38:03.506007+0000 mgr.a (mgr.14152) 250 : cluster [DBG] pgmap v154: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:05.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:04 vm09 bash[32072]: cluster 2026-03-31T11:38:03.506007+0000 mgr.a (mgr.14152) 250 : cluster [DBG] pgmap v154: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:05.084 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:04 vm09 bash[32072]: cluster 2026-03-31T11:38:03.506007+0000 mgr.a (mgr.14152) 250 : cluster [DBG] pgmap v154: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.906256+0000 mon.a (mon.0) 68 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.906256+0000 mon.a (mon.0) 68 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.907480+0000 mon.a (mon.0) 69 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.907480+0000 mon.a (mon.0) 69 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.907921+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.907921+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.911749+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.911749+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.913214+0000 mon.a (mon.0) 72 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:04.913214+0000 mon.a (mon.0) 72 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:05.007649+0000 mon.a (mon.0) 73 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:05.007649+0000 mon.a (mon.0) 73 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:05.011061+0000 mon.a (mon.0) 74 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:05.011061+0000 mon.a (mon.0) 74 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:05.014509+0000 mon.a (mon.0) 75 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:05.899 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:05 vm09 bash[32072]: audit 2026-03-31T11:38:05.014509+0000 mon.a (mon.0) 75 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.052 INFO:teuthology.orchestra.run.vm06.stderr:+ grep '\-\-cpus=2' /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/unit.run 2026-03-31T11:38:06.053 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-mon-a --pids-limit=0 --cpus=2 -e CONTAINER_IMAGE=quay.ceph.io/ceph-ci/ceph@sha256:02c8d616f8a7af1a26efca44a51de7761356c5cb66c69c789e66f798c27c8072 -e NODE_NAME=vm06 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5:/var/run/ceph:z -v /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5:/var/log/ceph:z -v /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config:/etc/ceph/ceph.conf:z quay.ceph.io/ceph-ci/ceph@sha256:02c8d616f8a7af1a26efca44a51de7761356c5cb66c69c789e66f798c27c8072 -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true '--default-log-stderr-prefix=debug ' --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 2026-03-31T11:38:06.053 INFO:teuthology.orchestra.run.vm06.stderr:+ grep '\-\-debug_ms 10' /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/unit.run 2026-03-31T11:38:06.054 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-mon-a --pids-limit=0 --cpus=2 -e CONTAINER_IMAGE=quay.ceph.io/ceph-ci/ceph@sha256:02c8d616f8a7af1a26efca44a51de7761356c5cb66c69c789e66f798c27c8072 -e NODE_NAME=vm06 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5:/var/run/ceph:z -v /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5:/var/log/ceph:z -v /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config:/etc/ceph/ceph.conf:z quay.ceph.io/ceph-ci/ceph@sha256:02c8d616f8a7af1a26efca44a51de7761356c5cb66c69c789e66f798c27c8072 -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true '--default-log-stderr-prefix=debug ' --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 2026-03-31T11:38:06.054 INFO:teuthology.orchestra.run.vm06.stderr:+ grep testing_custom_containers /etc/cephadm_testing/testing.txt 2026-03-31T11:38:06.055 INFO:teuthology.orchestra.run.vm06.stdout:testing_custom_containers 2026-03-31T11:38:06.055 INFO:teuthology.orchestra.run.vm06.stderr:++ find /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5 -maxdepth 1 -type d -name 'container.bar.*' 2026-03-31T11:38:06.056 INFO:teuthology.orchestra.run.vm06.stderr:+ dir=/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/container.bar.vm06 2026-03-31T11:38:06.056 INFO:teuthology.orchestra.run.vm06.stderr:+ test -n /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/container.bar.vm06 2026-03-31T11:38:06.056 INFO:teuthology.orchestra.run.vm06.stderr:+ grep ok /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/container.bar.vm06/data/primary.txt 2026-03-31T11:38:06.056 INFO:teuthology.orchestra.run.vm06.stdout:ok 2026-03-31T11:38:06.056 INFO:teuthology.orchestra.run.vm06.stderr:+ grep from=centos /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/container.bar.vm06/data/from.txt 2026-03-31T11:38:06.057 INFO:teuthology.orchestra.run.vm06.stdout:from=centos 2026-03-31T11:38:06.057 INFO:teuthology.orchestra.run.vm06.stderr:+ test -s /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/container.bar.vm06/data/presized.dat 2026-03-31T11:38:06.058 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-31T11:38:06.061 INFO:tasks.cephadm:Teardown begin 2026-03-31T11:38:06.061 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.906256+0000 mon.a (mon.0) 68 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.906256+0000 mon.a (mon.0) 68 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.907480+0000 mon.a (mon.0) 69 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.907480+0000 mon.a (mon.0) 69 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.907921+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.907921+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.911749+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.911749+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.913214+0000 mon.a (mon.0) 72 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:04.913214+0000 mon.a (mon.0) 72 : audit [DBG] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:05.007649+0000 mon.a (mon.0) 73 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:05.007649+0000 mon.a (mon.0) 73 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:05.011061+0000 mon.a (mon.0) 74 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:05.011061+0000 mon.a (mon.0) 74 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:05.014509+0000 mon.a (mon.0) 75 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.061 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 31 11:38:05 vm06 bash[37427]: audit 2026-03-31T11:38:05.014509+0000 mon.a (mon.0) 75 : audit [INF] from='mgr.14152 192.168.123.106:0/2220613713' entity='mgr.a' 2026-03-31T11:38:06.068 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-31T11:38:06.076 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-31T11:38:06.076 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:5bb3278730741031382ca9c3dc9d221a942e06a2 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 -- ceph mgr module disable cephadm 2026-03-31T11:38:06.360 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/mon.a/config 2026-03-31T11:38:06.544 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-31T11:38:06.540+0000 7f3643512640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-31T11:38:06.544 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-31T11:38:06.540+0000 7f3643512640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-31T11:38:06.544 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-31T11:38:06.540+0000 7f3643512640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-31T11:38:06.544 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-31T11:38:06.540+0000 7f3643512640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-31T11:38:06.544 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-31T11:38:06.540+0000 7f3643512640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-31T11:38:06.544 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-31T11:38:06.540+0000 7f3643512640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-31T11:38:06.544 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-31T11:38:06.540+0000 7f3643512640 -1 monclient: keyring not found 2026-03-31T11:38:06.544 INFO:teuthology.orchestra.run.vm06.stderr:[errno 21] error connecting to the cluster 2026-03-31T11:38:06.589 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-31T11:38:06.589 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-31T11:38:06.589 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-31T11:38:06.592 DEBUG:teuthology.orchestra.run.vm09:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-31T11:38:06.595 INFO:tasks.cephadm:Stopping all daemons... 2026-03-31T11:38:06.595 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-31T11:38:06.595 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a 2026-03-31T11:38:06.810 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.a.service' 2026-03-31T11:38:06.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:06 vm09 bash[32072]: cluster 2026-03-31T11:38:05.506203+0000 mgr.a (mgr.14152) 251 : cluster [DBG] pgmap v155: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:06.835 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:06 vm09 bash[32072]: cluster 2026-03-31T11:38:05.506203+0000 mgr.a (mgr.14152) 251 : cluster [DBG] pgmap v155: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-31T11:38:06.837 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-31T11:38:06.837 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-31T11:38:06.837 INFO:tasks.cephadm.mon.b:Stopping mon.b... 2026-03-31T11:38:06.837 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.b 2026-03-31T11:38:07.118 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:06 vm09 systemd[1]: Stopping Ceph mon.b for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:38:07.118 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:06 vm09 bash[32072]: debug 2026-03-31T11:38:06.878+0000 7fe7b5368640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 (PID: 1) UID: 0 2026-03-31T11:38:07.118 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 31 11:38:06 vm09 bash[32072]: debug 2026-03-31T11:38:06.878+0000 7fe7b5368640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-31T11:38:07.175 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mon.b.service' 2026-03-31T11:38:07.188 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-31T11:38:07.188 INFO:tasks.cephadm.mon.b:Stopped mon.b 2026-03-31T11:38:07.188 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-31T11:38:07.188 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.a 2026-03-31T11:38:07.347 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.a.service' 2026-03-31T11:38:07.358 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-31T11:38:07.358 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-31T11:38:07.358 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-31T11:38:07.358 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.b 2026-03-31T11:38:07.432 INFO:journalctl@ceph.mgr.b.vm09.stdout:Mar 31 11:38:07 vm09 systemd[1]: Stopping Ceph mgr.b for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:38:07.488 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@mgr.b.service' 2026-03-31T11:38:07.499 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-31T11:38:07.499 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-31T11:38:07.499 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-31T11:38:07.499 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@osd.0 2026-03-31T11:38:07.815 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:38:07 vm06 systemd[1]: Stopping Ceph osd.0 for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:38:07.816 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:38:07 vm06 bash[28816]: debug 2026-03-31T11:38:07.540+0000 7f9cffac9640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-31T11:38:07.816 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:38:07 vm06 bash[28816]: debug 2026-03-31T11:38:07.540+0000 7f9cffac9640 -1 osd.0 16 *** Got signal Terminated *** 2026-03-31T11:38:07.816 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:38:07 vm06 bash[28816]: debug 2026-03-31T11:38:07.540+0000 7f9cffac9640 -1 osd.0 16 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-31T11:38:12.863 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 31 11:38:12 vm06 bash[39189]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-osd-0 2026-03-31T11:38:12.892 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@osd.0.service' 2026-03-31T11:38:12.917 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-31T11:38:12.917 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-31T11:38:12.917 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-31T11:38:12.917 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@osd.1 2026-03-31T11:38:13.335 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:38:12 vm09 systemd[1]: Stopping Ceph osd.1 for 48134446-2cf5-11f1-830d-4121aa8d72d5... 2026-03-31T11:38:13.335 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:38:12 vm09 bash[27160]: debug 2026-03-31T11:38:12.958+0000 7f03c071d640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-31T11:38:13.335 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:38:12 vm09 bash[27160]: debug 2026-03-31T11:38:12.958+0000 7f03c071d640 -1 osd.1 16 *** Got signal Terminated *** 2026-03-31T11:38:13.335 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:38:12 vm09 bash[27160]: debug 2026-03-31T11:38:12.958+0000 7f03c071d640 -1 osd.1 16 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-31T11:38:18.304 INFO:journalctl@ceph.osd.1.vm09.stdout:Mar 31 11:38:18 vm09 bash[33409]: ceph-48134446-2cf5-11f1-830d-4121aa8d72d5-osd-1 2026-03-31T11:38:18.358 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-48134446-2cf5-11f1-830d-4121aa8d72d5@osd.1.service' 2026-03-31T11:38:18.379 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-31T11:38:18.379 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-31T11:38:18.379 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 --force --keep-logs 2026-03-31T11:38:18.628 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:38:21.468 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 --force --keep-logs 2026-03-31T11:38:21.726 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:38:24.263 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-31T11:38:24.272 INFO:teuthology.orchestra.run.vm06.stderr:rm: cannot remove '/etc/ceph/ceph.client.admin.keyring': Is a directory 2026-03-31T11:38:24.273 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-31T11:38:24.273 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-31T11:38:24.280 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-31T11:38:24.280 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/crash to /archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300/remote/vm06/crash 2026-03-31T11:38:24.280 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/crash -- . 2026-03-31T11:38:24.324 INFO:teuthology.orchestra.run.vm06.stderr:tar: /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/crash: Cannot open: No such file or directory 2026-03-31T11:38:24.324 INFO:teuthology.orchestra.run.vm06.stderr:tar: Error is not recoverable: exiting now 2026-03-31T11:38:24.324 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/crash to /archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300/remote/vm09/crash 2026-03-31T11:38:24.325 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/crash -- . 2026-03-31T11:38:24.333 INFO:teuthology.orchestra.run.vm09.stderr:tar: /var/lib/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/crash: Cannot open: No such file or directory 2026-03-31T11:38:24.333 INFO:teuthology.orchestra.run.vm09.stderr:tar: Error is not recoverable: exiting now 2026-03-31T11:38:24.333 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-31T11:38:24.333 DEBUG:teuthology.orchestra.run.vm06:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-31T11:38:24.378 INFO:tasks.cephadm:Compressing logs... 2026-03-31T11:38:24.378 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-31T11:38:24.421 DEBUG:teuthology.orchestra.run.vm09:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-31T11:38:24.427 INFO:teuthology.orchestra.run.vm06.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-31T11:38:24.428 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.audit.log 2026-03-31T11:38:24.428 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.log 2026-03-31T11:38:24.429 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.audit.log: gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-osd.0.log 2026-03-31T11:38:24.429 INFO:teuthology.orchestra.run.vm09.stderr:find: gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.audit.log 2026-03-31T11:38:24.429 INFO:teuthology.orchestra.run.vm09.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.log: gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.cephadm.log 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm06.stderr: 90.3% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.audit.log.gz 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm06.stderr: 86.7%/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-osd.0.log: -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.log.gz 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-volume.log 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mgr.a.log 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mon.b.log 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.cephadm.log: /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mon.a.log 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm06.stderr: 80.3% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.cephadm.log.gz 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.audit.log: gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mgr.b.log 2026-03-31T11:38:24.430 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-31T11:38:24.431 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mon.b.log: gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.log 2026-03-31T11:38:24.431 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mgr.b.log: 90.7% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.audit.log.gz 2026-03-31T11:38:24.431 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.cephadm.log 2026-03-31T11:38:24.431 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-volume.log 2026-03-31T11:38:24.432 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.log: /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.cephadm.log: 79.4% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.cephadm.log.gz 86.8% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph.log.gz 2026-03-31T11:38:24.432 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-31T11:38:24.432 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-osd.1.log 2026-03-31T11:38:24.432 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-31T11:38:24.433 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-volume.log: /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-osd.1.log: 90.9% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mgr.b.log.gz 2026-03-31T11:38:24.435 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/cephadm.log: 86.0% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-31T11:38:24.436 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mon.a.log: 96.1% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-volume.log.gz 2026-03-31T11:38:24.438 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/cephadm.log: 87.8% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-31T11:38:24.441 INFO:teuthology.orchestra.run.vm06.stderr: 94.1% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-osd.0.log.gz 2026-03-31T11:38:24.441 INFO:teuthology.orchestra.run.vm09.stderr: 96.2% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-volume.log.gz 2026-03-31T11:38:24.447 INFO:teuthology.orchestra.run.vm09.stderr: 94.0% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-osd.1.log.gz 2026-03-31T11:38:24.474 INFO:teuthology.orchestra.run.vm06.stderr: 88.4% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mgr.a.log.gz 2026-03-31T11:38:24.477 INFO:teuthology.orchestra.run.vm09.stderr: 93.8% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mon.b.log.gz 2026-03-31T11:38:24.478 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-31T11:38:24.478 INFO:teuthology.orchestra.run.vm09.stderr:real 0m0.054s 2026-03-31T11:38:24.478 INFO:teuthology.orchestra.run.vm09.stderr:user 0m0.084s 2026-03-31T11:38:24.478 INFO:teuthology.orchestra.run.vm09.stderr:sys 0m0.004s 2026-03-31T11:38:24.555 INFO:teuthology.orchestra.run.vm06.stderr: 91.9% -- replaced with /var/log/ceph/48134446-2cf5-11f1-830d-4121aa8d72d5/ceph-mon.a.log.gz 2026-03-31T11:38:24.557 INFO:teuthology.orchestra.run.vm06.stderr: 2026-03-31T11:38:24.557 INFO:teuthology.orchestra.run.vm06.stderr:real 0m0.135s 2026-03-31T11:38:24.557 INFO:teuthology.orchestra.run.vm06.stderr:user 0m0.173s 2026-03-31T11:38:24.557 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m0.031s 2026-03-31T11:38:24.557 INFO:tasks.cephadm:Archiving logs... 2026-03-31T11:38:24.557 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/log/ceph to /archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300/remote/vm06/log 2026-03-31T11:38:24.557 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-31T11:38:24.622 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/log/ceph to /archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300/remote/vm09/log 2026-03-31T11:38:24.622 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-31T11:38:24.634 INFO:tasks.cephadm:Removing cluster... 2026-03-31T11:38:24.634 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 --force 2026-03-31T11:38:24.908 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:38:25.138 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 48134446-2cf5-11f1-830d-4121aa8d72d5 --force 2026-03-31T11:38:25.379 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 48134446-2cf5-11f1-830d-4121aa8d72d5 2026-03-31T11:38:25.624 INFO:tasks.cephadm:Removing cephadm ... 2026-03-31T11:38:25.624 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-31T11:38:25.628 DEBUG:teuthology.orchestra.run.vm09:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-31T11:38:25.631 INFO:tasks.cephadm:Teardown complete 2026-03-31T11:38:25.631 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-31T11:38:25.633 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-31T11:38:25.633 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-31T11:38:25.677 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-31T11:38:25.691 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-31T11:38:25.691 DEBUG:teuthology.orchestra.run.vm06:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-31T11:38:25.696 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-31T11:38:25.696 DEBUG:teuthology.orchestra.run.vm09:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-31T11:38:25.756 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:25.758 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:25.924 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:25.924 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:25.927 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:25.927 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:26.055 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:26.056 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:38:26.056 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-31T11:38:26.056 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:26.063 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:26.064 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:38:26.064 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-31T11:38:26.064 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:26.068 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:26.069 INFO:teuthology.orchestra.run.vm09.stdout: ceph* 2026-03-31T11:38:26.075 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:26.075 INFO:teuthology.orchestra.run.vm06.stdout: ceph* 2026-03-31T11:38:26.229 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 1 to remove and 49 not upgraded. 2026-03-31T11:38:26.229 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-31T11:38:26.236 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 1 to remove and 49 not upgraded. 2026-03-31T11:38:26.236 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-31T11:38:26.268 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126150 files and directories currently installed.) 2026-03-31T11:38:26.270 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:26.272 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126150 files and directories currently installed.) 2026-03-31T11:38:26.274 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:27.299 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:27.315 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:27.333 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:27.350 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:27.504 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:27.505 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:27.524 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:27.524 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:27.638 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:27.638 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:38:27.638 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-31T11:38:27.638 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:27.649 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:27.650 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm* cephadm* 2026-03-31T11:38:27.670 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:27.670 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:38:27.670 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-31T11:38:27.670 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:27.679 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:27.680 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm* cephadm* 2026-03-31T11:38:27.812 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 2 to remove and 49 not upgraded. 2026-03-31T11:38:27.812 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 2177 kB disk space will be freed. 2026-03-31T11:38:27.843 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 2 to remove and 49 not upgraded. 2026-03-31T11:38:27.843 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 2177 kB disk space will be freed. 2026-03-31T11:38:27.850 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126148 files and directories currently installed.) 2026-03-31T11:38:27.852 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:27.869 INFO:teuthology.orchestra.run.vm09.stdout:Removing cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:27.881 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126148 files and directories currently installed.) 2026-03-31T11:38:27.883 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-mgr-cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:27.896 INFO:teuthology.orchestra.run.vm09.stdout:Looking for files to backup/remove ... 2026-03-31T11:38:27.898 INFO:teuthology.orchestra.run.vm09.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-31T11:38:27.900 INFO:teuthology.orchestra.run.vm09.stdout:Removing user `cephadm' ... 2026-03-31T11:38:27.900 INFO:teuthology.orchestra.run.vm09.stdout:Warning: group `nogroup' has no more members. 2026-03-31T11:38:27.901 INFO:teuthology.orchestra.run.vm06.stdout:Removing cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:27.910 INFO:teuthology.orchestra.run.vm09.stdout:Done. 2026-03-31T11:38:27.929 INFO:teuthology.orchestra.run.vm06.stdout:Looking for files to backup/remove ... 2026-03-31T11:38:27.930 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:27.931 INFO:teuthology.orchestra.run.vm06.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-31T11:38:27.933 INFO:teuthology.orchestra.run.vm06.stdout:Removing user `cephadm' ... 2026-03-31T11:38:27.933 INFO:teuthology.orchestra.run.vm06.stdout:Warning: group `nogroup' has no more members. 2026-03-31T11:38:27.944 INFO:teuthology.orchestra.run.vm06.stdout:Done. 2026-03-31T11:38:27.965 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:28.022 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126062 files and directories currently installed.) 2026-03-31T11:38:28.024 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:28.053 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126062 files and directories currently installed.) 2026-03-31T11:38:28.055 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for cephadm (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:29.044 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:29.076 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:29.080 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:29.110 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:29.248 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:29.248 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:29.278 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:29.278 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:29.371 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:29.372 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:38:29.372 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-31T11:38:29.372 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:29.382 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:29.383 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds* 2026-03-31T11:38:29.421 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:29.421 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-31T11:38:29.421 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-31T11:38:29.421 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:29.429 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:29.429 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds* 2026-03-31T11:38:29.545 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 1 to remove and 49 not upgraded. 2026-03-31T11:38:29.545 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 6851 kB disk space will be freed. 2026-03-31T11:38:29.584 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126062 files and directories currently installed.) 2026-03-31T11:38:29.587 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mds (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:29.589 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 1 to remove and 49 not upgraded. 2026-03-31T11:38:29.589 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 6851 kB disk space will be freed. 2026-03-31T11:38:29.627 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126062 files and directories currently installed.) 2026-03-31T11:38:29.629 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-mds (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:30.018 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:30.020 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:30.117 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126054 files and directories currently installed.) 2026-03-31T11:38:30.119 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for ceph-mds (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:30.121 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126054 files and directories currently installed.) 2026-03-31T11:38:30.124 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-mds (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:31.603 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:31.614 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:31.638 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:31.649 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:31.812 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:31.812 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:31.822 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:31.823 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:31.978 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh python3-cachetools python3-cheroot python3-cherrypy3 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-psutil 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: python3-threadpoolctl python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils sg3-utils-udev 2026-03-31T11:38:31.979 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:31.992 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:31.992 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-31T11:38:31.993 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-k8sevents* 2026-03-31T11:38:32.003 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:32.003 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh python3-cachetools python3-cheroot python3-cherrypy3 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-psutil 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: python3-threadpoolctl python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils sg3-utils-udev 2026-03-31T11:38:32.004 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:32.016 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:32.016 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-31T11:38:32.017 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents* 2026-03-31T11:38:32.156 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 4 to remove and 49 not upgraded. 2026-03-31T11:38:32.156 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 219 MB disk space will be freed. 2026-03-31T11:38:32.172 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 4 to remove and 49 not upgraded. 2026-03-31T11:38:32.172 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 219 MB disk space will be freed. 2026-03-31T11:38:32.192 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126054 files and directories currently installed.) 2026-03-31T11:38:32.194 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-k8sevents (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.205 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-diskprediction-local (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.208 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126054 files and directories currently installed.) 2026-03-31T11:38:32.210 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-mgr-k8sevents (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.220 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-mgr-diskprediction-local (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.231 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-dashboard (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.245 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-mgr-dashboard (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.298 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.316 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-mgr (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.744 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124271 files and directories currently installed.) 2026-03-31T11:38:32.747 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-mgr (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:32.779 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124271 files and directories currently installed.) 2026-03-31T11:38:32.781 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for ceph-mgr (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:34.093 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:34.125 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:34.128 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:34.161 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:34.282 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:34.283 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:34.319 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:34.319 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:34.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:34.399 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:34.399 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:34.409 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:34.409 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:34.447 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:34.455 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:34.455 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-31T11:38:34.563 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 7 to remove and 49 not upgraded. 2026-03-31T11:38:34.563 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 732 MB disk space will be freed. 2026-03-31T11:38:34.598 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124271 files and directories currently installed.) 2026-03-31T11:38:34.600 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-volume (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:34.606 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 7 to remove and 49 not upgraded. 2026-03-31T11:38:34.606 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 732 MB disk space will be freed. 2026-03-31T11:38:34.641 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124271 files and directories currently installed.) 2026-03-31T11:38:34.643 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-volume (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:34.655 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-osd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:34.698 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-osd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:35.059 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:35.081 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-mon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:35.442 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-base (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:35.450 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-base (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:35.813 INFO:teuthology.orchestra.run.vm06.stdout:Removing radosgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:35.891 INFO:teuthology.orchestra.run.vm09.stdout:Removing radosgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:36.192 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-test (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:36.248 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:36.291 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-test (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:36.352 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:36.618 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:36.648 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:38:36.709 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123776 files and directories currently installed.) 2026-03-31T11:38:36.711 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for radosgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:36.786 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:36.816 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:38:36.878 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123776 files and directories currently installed.) 2026-03-31T11:38:36.880 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for radosgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:37.227 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for ceph-mon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:37.468 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-mon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:37.584 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for ceph-base (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:37.863 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-base (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:37.948 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:38.259 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:38.334 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for ceph-osd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:38.661 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-osd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:39.642 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:39.676 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:39.830 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:39.831 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:39.935 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:39.936 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:39.944 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:39.944 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse* 2026-03-31T11:38:40.006 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:40.039 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:40.095 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 1 to remove and 49 not upgraded. 2026-03-31T11:38:40.095 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 2932 kB disk space will be freed. 2026-03-31T11:38:40.129 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123764 files and directories currently installed.) 2026-03-31T11:38:40.131 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:40.193 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:40.193 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:40.294 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:40.294 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:40.294 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:40.295 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:40.303 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:40.303 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse* 2026-03-31T11:38:40.450 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 1 to remove and 49 not upgraded. 2026-03-31T11:38:40.450 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 2932 kB disk space will be freed. 2026-03-31T11:38:40.485 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123764 files and directories currently installed.) 2026-03-31T11:38:40.486 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:40.518 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:40.597 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123755 files and directories currently installed.) 2026-03-31T11:38:40.599 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for ceph-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:40.909 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:40.991 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123755 files and directories currently installed.) 2026-03-31T11:38:40.995 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:41.859 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:41.893 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:42.058 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:42.058 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:42.164 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:42.178 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:42.178 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:42.210 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:42.346 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:42.359 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:42.359 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:42.379 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:42.483 INFO:teuthology.orchestra.run.vm06.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:42.484 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:42.484 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:42.484 INFO:teuthology.orchestra.run.vm06.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:42.484 INFO:teuthology.orchestra.run.vm06.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:42.484 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:42.484 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:42.502 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:42.502 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:42.534 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:42.560 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:42.560 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:42.661 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:42.676 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:42.676 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:42.701 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:42.702 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:42.708 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout:Package 'radosgw' is not installed, so not removed 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:42.809 INFO:teuthology.orchestra.run.vm06.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:42.810 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:42.810 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:42.810 INFO:teuthology.orchestra.run.vm06.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:42.810 INFO:teuthology.orchestra.run.vm06.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:42.810 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:42.810 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:42.824 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:42.824 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:42.856 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:42.877 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:42.878 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:42.977 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:42.992 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:42.992 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:43.024 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:43.025 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:43.026 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph nvme-cli 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:43.134 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:43.135 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:43.135 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:43.135 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:43.135 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:43.135 INFO:teuthology.orchestra.run.vm06.stdout: socat xmlstarlet 2026-03-31T11:38:43.135 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:43.143 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:43.143 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-31T11:38:43.188 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:43.189 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:43.293 INFO:teuthology.orchestra.run.vm09.stdout:Package 'radosgw' is not installed, so not removed 2026-03-31T11:38:43.293 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:43.293 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:43.293 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:43.293 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-31T11:38:43.293 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh python3-cachetools python3-ceph-common python3-cheroot 2026-03-31T11:38:43.293 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-31T11:38:43.294 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-31T11:38:43.294 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-31T11:38:43.294 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-31T11:38:43.294 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-31T11:38:43.294 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-31T11:38:43.294 INFO:teuthology.orchestra.run.vm09.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-31T11:38:43.294 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:43.294 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:43.300 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 3 to remove and 49 not upgraded. 2026-03-31T11:38:43.301 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 2086 kB disk space will be freed. 2026-03-31T11:38:43.309 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:43.309 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:43.334 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123755 files and directories currently installed.) 2026-03-31T11:38:43.336 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-cephfs (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:43.342 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:43.348 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-rgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:43.358 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-rados (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:43.517 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:43.517 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph nvme-cli 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout: socat xmlstarlet 2026-03-31T11:38:43.627 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:43.638 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:43.638 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-31T11:38:43.805 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 3 to remove and 49 not upgraded. 2026-03-31T11:38:43.805 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 2086 kB disk space will be freed. 2026-03-31T11:38:43.844 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123755 files and directories currently installed.) 2026-03-31T11:38:43.846 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-cephfs (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:43.858 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-rgw (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:43.868 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-rados (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:44.452 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:44.485 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:44.641 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:44.642 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:44.766 INFO:teuthology.orchestra.run.vm06.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph nvme-cli 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout: socat xmlstarlet 2026-03-31T11:38:44.767 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:44.782 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:44.783 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:44.816 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:44.846 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:44.879 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:44.996 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:44.996 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:45.058 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:45.059 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph nvme-cli 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout: socat xmlstarlet 2026-03-31T11:38:45.118 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:45.138 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:45.138 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:45.170 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:45.203 INFO:teuthology.orchestra.run.vm09.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph nvme-cli 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout: socat xmlstarlet 2026-03-31T11:38:45.204 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:45.220 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:45.221 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:45.253 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:45.343 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:45.343 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:45.434 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:45.435 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:45.489 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph nvme-cli 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:45.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:45.491 INFO:teuthology.orchestra.run.vm06.stdout: socat xmlstarlet 2026-03-31T11:38:45.491 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:45.505 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:45.505 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd* 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph nvme-cli 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:45.577 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:45.578 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:45.578 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:45.578 INFO:teuthology.orchestra.run.vm09.stdout: socat xmlstarlet 2026-03-31T11:38:45.578 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:45.594 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:45.594 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:45.627 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:45.677 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 1 to remove and 49 not upgraded. 2026-03-31T11:38:45.677 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 1205 kB disk space will be freed. 2026-03-31T11:38:45.717 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123731 files and directories currently installed.) 2026-03-31T11:38:45.720 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-rbd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:45.830 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:45.830 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:45.984 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:45.984 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:45.984 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-31T11:38:45.984 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph nvme-cli 2026-03-31T11:38:45.984 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:45.984 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout: socat xmlstarlet 2026-03-31T11:38:45.985 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:45.994 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:45.994 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd* 2026-03-31T11:38:46.149 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 1 to remove and 49 not upgraded. 2026-03-31T11:38:46.149 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 1205 kB disk space will be freed. 2026-03-31T11:38:46.186 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123731 files and directories currently installed.) 2026-03-31T11:38:46.188 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-rbd (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:46.742 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:46.777 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:46.931 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:46.932 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:47.041 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout: socat xmlstarlet 2026-03-31T11:38:47.042 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:47.051 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:47.051 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon* libcephfs-dev* libcephfs2* 2026-03-31T11:38:47.164 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:47.197 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:47.203 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 3 to remove and 49 not upgraded. 2026-03-31T11:38:47.203 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 2851 kB disk space will be freed. 2026-03-31T11:38:47.239 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123723 files and directories currently installed.) 2026-03-31T11:38:47.241 INFO:teuthology.orchestra.run.vm06.stdout:Removing libcephfs-daemon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:47.252 INFO:teuthology.orchestra.run.vm06.stdout:Removing libcephfs-dev (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:47.262 INFO:teuthology.orchestra.run.vm06.stdout:Removing libcephfs2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:47.285 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:38:47.352 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:47.352 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:47.471 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:47.471 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:47.471 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 2026-03-31T11:38:47.471 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph 2026-03-31T11:38:47.471 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:47.471 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout: socat xmlstarlet 2026-03-31T11:38:47.472 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:47.481 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:47.482 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-daemon* libcephfs-dev* libcephfs2* 2026-03-31T11:38:47.635 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 3 to remove and 49 not upgraded. 2026-03-31T11:38:47.635 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 2851 kB disk space will be freed. 2026-03-31T11:38:47.670 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123723 files and directories currently installed.) 2026-03-31T11:38:47.672 INFO:teuthology.orchestra.run.vm09.stdout:Removing libcephfs-daemon (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:47.683 INFO:teuthology.orchestra.run.vm09.stdout:Removing libcephfs-dev (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:47.694 INFO:teuthology.orchestra.run.vm09.stdout:Removing libcephfs2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:47.717 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:38:48.288 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:48.321 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:48.480 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:48.481 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout: socat xmlstarlet 2026-03-31T11:38:48.590 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:48.605 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:48.605 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:48.636 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:48.696 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:48.728 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:48.797 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:48.798 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:48.889 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:48.890 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libcephfs-proxy2 libdouble-conversion3 libfuse2 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: libgfapi0 libgfrpc0 libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: libnbd0 liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: libqt5network5 librdkafka1 libsgutils2-2 libthrift-0.16.0 nvme-cli 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n sg3-utils 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:48.902 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:48.913 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:48.913 INFO:teuthology.orchestra.run.vm06.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-31T11:38:48.913 INFO:teuthology.orchestra.run.vm06.stdout: qemu-block-extra* rbd-fuse* 2026-03-31T11:38:49.001 INFO:teuthology.orchestra.run.vm09.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-31T11:38:49.001 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:49.001 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:49.001 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 2026-03-31T11:38:49.001 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 librgw2 libsgutils2-2 libsqlite3-mod-ceph 2026-03-31T11:38:49.001 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:49.001 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout: socat xmlstarlet 2026-03-31T11:38:49.002 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:49.017 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:49.017 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:49.049 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:49.065 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 7 to remove and 49 not upgraded. 2026-03-31T11:38:49.065 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 59.2 MB disk space will be freed. 2026-03-31T11:38:49.099 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123701 files and directories currently installed.) 2026-03-31T11:38:49.101 INFO:teuthology.orchestra.run.vm06.stdout:Removing rbd-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.113 INFO:teuthology.orchestra.run.vm06.stdout:Removing libsqlite3-mod-ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.124 INFO:teuthology.orchestra.run.vm06.stdout:Removing libradosstriper1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.135 INFO:teuthology.orchestra.run.vm06.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-31T11:38:49.198 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:49.198 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:49.300 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:49.300 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:49.300 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libcephfs-proxy2 libdouble-conversion3 libfuse2 2026-03-31T11:38:49.300 INFO:teuthology.orchestra.run.vm09.stdout: libgfapi0 libgfrpc0 libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 2026-03-31T11:38:49.300 INFO:teuthology.orchestra.run.vm09.stdout: libnbd0 liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-31T11:38:49.300 INFO:teuthology.orchestra.run.vm09.stdout: libqt5network5 librdkafka1 libsgutils2-2 libthrift-0.16.0 nvme-cli 2026-03-31T11:38:49.300 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n sg3-utils 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:49.301 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:49.309 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:49.309 INFO:teuthology.orchestra.run.vm09.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-31T11:38:49.309 INFO:teuthology.orchestra.run.vm09.stdout: qemu-block-extra* rbd-fuse* 2026-03-31T11:38:49.459 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 7 to remove and 49 not upgraded. 2026-03-31T11:38:49.460 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 59.2 MB disk space will be freed. 2026-03-31T11:38:49.495 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123701 files and directories currently installed.) 2026-03-31T11:38:49.497 INFO:teuthology.orchestra.run.vm09.stdout:Removing rbd-fuse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.511 INFO:teuthology.orchestra.run.vm09.stdout:Removing libsqlite3-mod-ceph (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.521 INFO:teuthology.orchestra.run.vm06.stdout:Removing librbd1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.522 INFO:teuthology.orchestra.run.vm09.stdout:Removing libradosstriper1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.533 INFO:teuthology.orchestra.run.vm06.stdout:Removing librgw2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.533 INFO:teuthology.orchestra.run.vm09.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-31T11:38:49.546 INFO:teuthology.orchestra.run.vm06.stdout:Removing librados2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.570 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:49.599 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:38:49.662 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123650 files and directories currently installed.) 2026-03-31T11:38:49.664 INFO:teuthology.orchestra.run.vm06.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-31T11:38:49.911 INFO:teuthology.orchestra.run.vm09.stdout:Removing librbd1 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.923 INFO:teuthology.orchestra.run.vm09.stdout:Removing librgw2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.937 INFO:teuthology.orchestra.run.vm09.stdout:Removing librados2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:49.961 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:49.995 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:38:50.063 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123650 files and directories currently installed.) 2026-03-31T11:38:50.065 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-31T11:38:50.952 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:50.986 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:51.145 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:51.145 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout:Package 'librbd1' is not installed, so not removed 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libcephfs-proxy2 libdouble-conversion3 libfuse2 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: libgfapi0 libgfrpc0 libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: libnbd0 liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: libqt5network5 librdkafka1 libsgutils2-2 libthrift-0.16.0 nvme-cli 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n sg3-utils 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:51.248 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:51.266 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:51.266 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:51.299 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:51.446 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:51.446 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:51.463 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:51.497 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:51.556 INFO:teuthology.orchestra.run.vm06.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libcephfs-proxy2 libdouble-conversion3 libfuse2 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: libgfapi0 libgfrpc0 libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: libnbd0 liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: libqt5network5 librdkafka1 libsgutils2-2 libthrift-0.16.0 nvme-cli 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n sg3-utils 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:51.557 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:51.572 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:51.572 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:51.573 DEBUG:teuthology.orchestra.run.vm06:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-31T11:38:51.628 DEBUG:teuthology.orchestra.run.vm06:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-31T11:38:51.659 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:51.660 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:51.702 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout:Package 'librbd1' is not installed, so not removed 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libcephfs-proxy2 libdouble-conversion3 libfuse2 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: libgfapi0 libgfrpc0 libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: libnbd0 liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: libqt5network5 librdkafka1 libsgutils2-2 libthrift-0.16.0 nvme-cli 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n sg3-utils 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:51.765 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:51.781 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:51.781 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:51.813 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:51.869 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-31T11:38:51.869 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-31T11:38:51.977 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:51.977 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be REMOVED: 2026-03-31T11:38:51.977 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:51.977 INFO:teuthology.orchestra.run.vm06.stdout: libboost-thread1.74.0 libcephfs-proxy2 libdouble-conversion3 libfuse2 2026-03-31T11:38:51.977 INFO:teuthology.orchestra.run.vm06.stdout: libgfapi0 libgfrpc0 libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 2026-03-31T11:38:51.977 INFO:teuthology.orchestra.run.vm06.stdout: libnbd0 liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: libqt5network5 librdkafka1 libsgutils2-2 libthrift-0.16.0 nvme-cli 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n sg3-utils 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm06.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:51.978 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:52.084 INFO:teuthology.orchestra.run.vm09.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-31T11:38:52.084 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-31T11:38:52.084 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:52.084 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libcephfs-proxy2 libdouble-conversion3 libfuse2 2026-03-31T11:38:52.084 INFO:teuthology.orchestra.run.vm09.stdout: libgfapi0 libgfrpc0 libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: libnbd0 liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: libqt5network5 librdkafka1 libsgutils2-2 libthrift-0.16.0 nvme-cli 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n sg3-utils 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:52.085 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-31T11:38:52.099 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 49 not upgraded. 2026-03-31T11:38:52.100 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:52.101 DEBUG:teuthology.orchestra.run.vm09:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-31T11:38:52.131 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 0 newly installed, 64 to remove and 49 not upgraded. 2026-03-31T11:38:52.131 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 96.8 MB disk space will be freed. 2026-03-31T11:38:52.154 DEBUG:teuthology.orchestra.run.vm09:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-31T11:38:52.165 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123650 files and directories currently installed.) 2026-03-31T11:38:52.166 INFO:teuthology.orchestra.run.vm06.stdout:Removing ceph-mgr-modules-core (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:52.179 INFO:teuthology.orchestra.run.vm06.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-31T11:38:52.188 INFO:teuthology.orchestra.run.vm06.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-31T11:38:52.198 INFO:teuthology.orchestra.run.vm06.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-31T11:38:52.207 INFO:teuthology.orchestra.run.vm06.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-31T11:38:52.216 INFO:teuthology.orchestra.run.vm06.stdout:Removing libcephfs-proxy2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:52.226 INFO:teuthology.orchestra.run.vm06.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-31T11:38:52.230 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:38:52.236 INFO:teuthology.orchestra.run.vm06.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:38:52.249 INFO:teuthology.orchestra.run.vm06.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:38:52.259 INFO:teuthology.orchestra.run.vm06.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:38:52.278 INFO:teuthology.orchestra.run.vm06.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-31T11:38:52.287 INFO:teuthology.orchestra.run.vm06.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-31T11:38:52.297 INFO:teuthology.orchestra.run.vm06.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-31T11:38:52.306 INFO:teuthology.orchestra.run.vm06.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-31T11:38:52.316 INFO:teuthology.orchestra.run.vm06.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-31T11:38:52.325 INFO:teuthology.orchestra.run.vm06.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-31T11:38:52.335 INFO:teuthology.orchestra.run.vm06.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-31T11:38:52.344 INFO:teuthology.orchestra.run.vm06.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-31T11:38:52.354 INFO:teuthology.orchestra.run.vm06.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-31T11:38:52.364 INFO:teuthology.orchestra.run.vm06.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-31T11:38:52.374 INFO:teuthology.orchestra.run.vm06.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-31T11:38:52.384 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-31T11:38:52.384 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-31T11:38:52.385 INFO:teuthology.orchestra.run.vm06.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-31T11:38:52.395 INFO:teuthology.orchestra.run.vm06.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-31T11:38:52.405 INFO:teuthology.orchestra.run.vm06.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-31T11:38:52.416 INFO:teuthology.orchestra.run.vm06.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-31T11:38:52.426 INFO:teuthology.orchestra.run.vm06.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-31T11:38:52.433 INFO:teuthology.orchestra.run.vm06.stdout:update-initramfs: deferring update (trigger activated) 2026-03-31T11:38:52.445 INFO:teuthology.orchestra.run.vm06.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-31T11:38:52.460 INFO:teuthology.orchestra.run.vm06.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-31T11:38:52.470 INFO:teuthology.orchestra.run.vm06.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libcephfs-proxy2 libdouble-conversion3 libfuse2 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: libgfapi0 libgfrpc0 libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: libnbd0 liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: libqt5network5 librdkafka1 libsgutils2-2 libthrift-0.16.0 nvme-cli 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cheroot python3-cherrypy3 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth python3-jaraco.classes python3-jaraco.collections 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes python3-natsort python3-portend python3-prettytable 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-wcwidth python3-webob 2026-03-31T11:38:52.492 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n sg3-utils 2026-03-31T11:38:52.493 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-31T11:38:52.645 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 64 to remove and 49 not upgraded. 2026-03-31T11:38:52.645 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 96.8 MB disk space will be freed. 2026-03-31T11:38:52.685 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123650 files and directories currently installed.) 2026-03-31T11:38:52.687 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-modules-core (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:52.702 INFO:teuthology.orchestra.run.vm09.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-31T11:38:52.713 INFO:teuthology.orchestra.run.vm09.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-31T11:38:52.724 INFO:teuthology.orchestra.run.vm09.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-31T11:38:52.735 INFO:teuthology.orchestra.run.vm09.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-31T11:38:52.745 INFO:teuthology.orchestra.run.vm09.stdout:Removing libcephfs-proxy2 (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:52.756 INFO:teuthology.orchestra.run.vm09.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-31T11:38:52.767 INFO:teuthology.orchestra.run.vm09.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:38:52.777 INFO:teuthology.orchestra.run.vm09.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:38:52.788 INFO:teuthology.orchestra.run.vm09.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-31T11:38:52.807 INFO:teuthology.orchestra.run.vm09.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-31T11:38:52.817 INFO:teuthology.orchestra.run.vm09.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-31T11:38:52.827 INFO:teuthology.orchestra.run.vm09.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-31T11:38:52.837 INFO:teuthology.orchestra.run.vm09.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-31T11:38:52.847 INFO:teuthology.orchestra.run.vm09.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-31T11:38:52.859 INFO:teuthology.orchestra.run.vm09.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-31T11:38:52.859 INFO:teuthology.orchestra.run.vm06.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-31T11:38:52.870 INFO:teuthology.orchestra.run.vm09.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-31T11:38:52.873 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-31T11:38:52.881 INFO:teuthology.orchestra.run.vm09.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-31T11:38:52.891 INFO:teuthology.orchestra.run.vm09.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-31T11:38:52.903 INFO:teuthology.orchestra.run.vm09.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-31T11:38:52.913 INFO:teuthology.orchestra.run.vm09.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-31T11:38:52.923 INFO:teuthology.orchestra.run.vm09.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-31T11:38:52.928 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-31T11:38:52.933 INFO:teuthology.orchestra.run.vm09.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-31T11:38:52.944 INFO:teuthology.orchestra.run.vm09.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-31T11:38:52.955 INFO:teuthology.orchestra.run.vm09.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-31T11:38:52.965 INFO:teuthology.orchestra.run.vm09.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-31T11:38:52.972 INFO:teuthology.orchestra.run.vm09.stdout:update-initramfs: deferring update (trigger activated) 2026-03-31T11:38:52.982 INFO:teuthology.orchestra.run.vm09.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-31T11:38:52.997 INFO:teuthology.orchestra.run.vm09.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-31T11:38:53.008 INFO:teuthology.orchestra.run.vm09.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-31T11:38:53.183 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-31T11:38:53.236 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-31T11:38:53.284 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-ceph-argparse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:53.330 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:53.381 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-31T11:38:53.398 INFO:teuthology.orchestra.run.vm09.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-31T11:38:53.411 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-31T11:38:53.444 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.2) ... 2026-03-31T11:38:53.464 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-31T11:38:53.492 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-31T11:38:53.536 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-31T11:38:53.580 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-31T11:38:53.629 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-31T11:38:53.673 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-31T11:38:53.712 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-31T11:38:53.718 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-31T11:38:53.761 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-31T11:38:53.763 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-31T11:38:53.808 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-ceph-argparse (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:53.854 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-ceph-common (20.2.0-721-g5bb32787-1jammy) ... 2026-03-31T11:38:53.874 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-31T11:38:53.905 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-31T11:38:53.932 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-31T11:38:53.972 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.2) ... 2026-03-31T11:38:53.980 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-31T11:38:54.019 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-31T11:38:54.026 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-31T11:38:54.064 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-31T11:38:54.074 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-31T11:38:54.109 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-31T11:38:54.121 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-31T11:38:54.154 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-31T11:38:54.167 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-31T11:38:54.200 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-31T11:38:54.216 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-rsa (4.8-1) ... 2026-03-31T11:38:54.246 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-31T11:38:54.264 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-31T11:38:54.296 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-31T11:38:54.317 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-31T11:38:54.330 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-31T11:38:54.376 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-31T11:38:54.407 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-31T11:38:54.424 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-31T11:38:54.467 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-31T11:38:54.474 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-31T11:38:54.515 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-31T11:38:54.525 INFO:teuthology.orchestra.run.vm06.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-31T11:38:54.562 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-31T11:38:54.571 INFO:teuthology.orchestra.run.vm06.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-31T11:38:54.589 INFO:teuthology.orchestra.run.vm06.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-31T11:38:54.610 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-31T11:38:54.657 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-31T11:38:54.702 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-31T11:38:54.771 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-rsa (4.8-1) ... 2026-03-31T11:38:54.819 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-31T11:38:54.870 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-31T11:38:54.883 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-31T11:38:54.931 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-31T11:38:54.975 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-31T11:38:55.008 INFO:teuthology.orchestra.run.vm06.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-31T11:38:55.019 INFO:teuthology.orchestra.run.vm06.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-31T11:38:55.021 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-31T11:38:55.053 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:38:55.063 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:55.070 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-31T11:38:55.105 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-31T11:38:55.115 INFO:teuthology.orchestra.run.vm09.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-31T11:38:55.120 INFO:teuthology.orchestra.run.vm06.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-171-generic 2026-03-31T11:38:55.134 INFO:teuthology.orchestra.run.vm09.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-31T11:38:55.541 INFO:teuthology.orchestra.run.vm09.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-31T11:38:55.552 INFO:teuthology.orchestra.run.vm09.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-31T11:38:55.616 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-31T11:38:55.627 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-31T11:38:55.672 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-31T11:38:55.689 INFO:teuthology.orchestra.run.vm09.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-171-generic 2026-03-31T11:38:59.491 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:59.494 DEBUG:teuthology.parallel:result is None 2026-03-31T11:38:59.916 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-31T11:38:59.918 DEBUG:teuthology.parallel:result is None 2026-03-31T11:38:59.918 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm06.local 2026-03-31T11:38:59.918 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm09.local 2026-03-31T11:38:59.918 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-31T11:38:59.919 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-31T11:38:59.926 DEBUG:teuthology.orchestra.run.vm06:> sudo apt-get update 2026-03-31T11:38:59.968 DEBUG:teuthology.orchestra.run.vm09:> sudo apt-get update 2026-03-31T11:39:00.075 INFO:teuthology.orchestra.run.vm06.stdout:Hit:1 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-31T11:39:00.100 INFO:teuthology.orchestra.run.vm06.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-31T11:39:00.130 INFO:teuthology.orchestra.run.vm06.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-31T11:39:00.134 INFO:teuthology.orchestra.run.vm09.stdout:Hit:1 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-31T11:39:00.134 INFO:teuthology.orchestra.run.vm09.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-31T11:39:00.165 INFO:teuthology.orchestra.run.vm09.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-31T11:39:00.200 INFO:teuthology.orchestra.run.vm09.stdout:Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-31T11:39:00.241 INFO:teuthology.orchestra.run.vm06.stdout:Hit:4 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-31T11:39:00.977 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-31T11:39:00.989 DEBUG:teuthology.parallel:result is None 2026-03-31T11:39:01.010 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-31T11:39:01.022 DEBUG:teuthology.parallel:result is None 2026-03-31T11:39:01.022 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-31T11:39:01.024 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-31T11:39:01.024 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-31T11:39:01.025 DEBUG:teuthology.orchestra.run.vm09:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout: remote refid st t when poll reach delay offset jitter 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================== 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:-ntp2.lwlcom.net .GPS. 1 u 62 64 177 30.903 +2.862 0.436 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:#pve2.h4x-gamers 79.133.44.142 2 u 58 64 177 25.040 -0.798 0.745 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:-sonne.floppy.or 237.17.204.95 2 u 54 64 177 28.779 +0.077 0.273 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:-ntp2.kernfusion 237.17.204.95 2 u 56 64 177 31.105 +0.416 0.645 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:+node-1.infogral 168.239.11.197 2 u 55 64 177 23.506 -1.047 1.148 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:-formularfetisch 131.188.3.223 2 u 54 64 177 25.044 -0.397 0.304 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:-ntp3.rrze.uni-e .PZFs. 1 u 58 64 177 26.036 -1.158 0.835 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:-static.81.54.25 131.188.3.222 2 u 57 64 177 25.074 -0.375 0.333 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:+ntp2.adminforge 131.188.3.220 2 u 54 64 177 24.990 -0.289 0.287 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:#v22025082392863 129.69.253.1 2 u 53 64 177 28.656 -3.262 0.282 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:#185.125.190.57 99.220.8.133 2 u 63 64 177 34.430 -2.040 0.361 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:*static.222.16.4 35.73.197.144 2 u 55 64 137 0.330 -1.096 1.022 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:-home.of.the.smi .BBgp. 1 u 53 64 177 38.073 +2.128 0.377 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:#alphyn.canonica 132.163.96.1 2 u 63 64 177 109.538 -1.304 0.392 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:#static.119.109. 148.251.235.164 3 u 52 64 177 23.637 -0.799 0.278 2026-03-31T11:39:01.420 INFO:teuthology.orchestra.run.vm09.stdout:#185.125.190.56 146.131.121.246 2 u 63 64 177 31.171 -2.429 2.157 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout: remote refid st t when poll reach delay offset jitter 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================== 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:-tick.infra.9rc. 46.38.244.94 3 u 54 64 177 28.251 -4.117 0.392 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:+ntp2.adminforge 131.188.3.220 2 u 55 64 177 25.043 -0.307 0.918 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:+formularfetisch 131.188.3.223 2 u 57 64 177 24.997 -0.748 0.634 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:#ntp3.rrze.uni-e .PZFs. 1 u 57 64 177 26.332 -1.234 0.372 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:+static.81.54.25 131.188.3.222 2 u 57 64 177 25.023 -0.426 0.212 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:-v22025082392863 129.69.253.1 2 u 52 64 177 28.230 -3.079 0.238 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:-ntp2.kernfusion 237.17.204.95 2 u 52 64 177 29.874 +0.678 1.368 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:+node-1.infogral 168.239.11.197 2 u 52 64 177 23.491 -0.247 0.340 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:+pve2.h4x-gamers 79.133.44.142 2 u 60 64 177 25.017 -0.427 0.217 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:#185.125.190.57 99.220.8.133 2 u - 64 377 34.293 -1.429 0.314 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:-home.of.the.smi .BBgp. 1 u 60 64 177 38.235 +1.854 0.346 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:*static.222.16.4 35.73.197.144 2 u 53 64 177 0.507 -0.264 0.702 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:#sonne.floppy.or 237.17.204.95 2 u 54 64 177 29.058 +0.351 0.296 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:#185.125.190.58 99.220.8.133 2 u 7 64 377 34.381 -1.378 0.353 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:#static.119.109. 148.251.235.164 3 u 55 64 177 23.632 -0.915 0.197 2026-03-31T11:39:01.421 INFO:teuthology.orchestra.run.vm06.stdout:#185.125.190.56 146.131.121.246 2 u 63 64 177 34.357 -2.145 0.445 2026-03-31T11:39:01.421 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-31T11:39:01.423 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-31T11:39:01.423 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-31T11:39:01.426 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-31T11:39:01.428 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-31T11:39:01.430 INFO:teuthology.task.internal:Duration was 608.642061 seconds 2026-03-31T11:39:01.430 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-31T11:39:01.432 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-31T11:39:01.432 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-31T11:39:01.433 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-31T11:39:01.453 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-31T11:39:01.453 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-03-31T11:39:01.453 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-31T11:39:01.499 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm09.local 2026-03-31T11:39:01.499 DEBUG:teuthology.orchestra.run.vm09:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-31T11:39:01.507 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-31T11:39:01.507 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-31T11:39:01.541 DEBUG:teuthology.orchestra.run.vm09:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-31T11:39:01.587 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-31T11:39:01.588 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-31T11:39:01.621 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-31T11:39:01.628 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-31T11:39:01.628 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-31T11:39:01.628 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-31T11:39:01.628 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-31T11:39:01.629 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-31T11:39:01.636 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-31T11:39:01.637 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-31T11:39:01.637 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gzgzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-31T11:39:01.637 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-31T11:39:01.637 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 88.5% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-31T11:39:01.637 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-31T11:39:01.645 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 88.7% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-31T11:39:01.646 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-31T11:39:01.648 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-31T11:39:01.649 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-31T11:39:01.688 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-31T11:39:01.697 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-31T11:39:01.699 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-31T11:39:01.729 DEBUG:teuthology.orchestra.run.vm09:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-31T11:39:01.734 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-03-31T11:39:01.743 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = core 2026-03-31T11:39:01.751 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-31T11:39:01.786 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-31T11:39:01.786 DEBUG:teuthology.orchestra.run.vm09:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-31T11:39:01.795 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-31T11:39:01.795 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-31T11:39:01.798 INFO:teuthology.task.internal:Transferring archived files... 2026-03-31T11:39:01.798 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300/remote/vm06 2026-03-31T11:39:01.798 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-31T11:39:01.835 DEBUG:teuthology.misc:Transferring archived files from vm09:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-31_11:18:10-rados-tentacle-none-default-vps/4300/remote/vm09 2026-03-31T11:39:01.835 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-31T11:39:01.844 INFO:teuthology.task.internal:Removing archive directory... 2026-03-31T11:39:01.844 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-31T11:39:01.877 DEBUG:teuthology.orchestra.run.vm09:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-31T11:39:01.888 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-31T11:39:01.890 INFO:teuthology.task.internal:Not uploading archives. 2026-03-31T11:39:01.890 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-31T11:39:01.893 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-31T11:39:01.893 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-31T11:39:01.921 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-31T11:39:01.923 INFO:teuthology.orchestra.run.vm06.stdout: 258079 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 31 11:39 /home/ubuntu/cephtest 2026-03-31T11:39:01.932 INFO:teuthology.orchestra.run.vm09.stdout: 258067 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 31 11:39 /home/ubuntu/cephtest 2026-03-31T11:39:01.933 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-31T11:39:01.938 INFO:teuthology.run:Summary data: description: rados/cephadm/workunits/{0-distro/ubuntu_22.04 agent/on mon_election/connectivity task/test_extra_daemon_features} duration: 608.6420607566833 flavor: default owner: kyr success: true 2026-03-31T11:39:01.938 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-31T11:39:01.955 INFO:teuthology.run:pass